searching-mlflow-docs
MLflow Documentation Search
Workflow
- Fetch
https://mlflow.org/docs/latest/llms.txtto find relevant page paths - Fetch the
.mdfile at the identified path - Present results with verbatim code examples
Step 1: Fetch llms.txt Index
WebFetch(
url: "https://mlflow.org/docs/latest/llms.txt",
prompt: "Find links or references to [TOPIC]. List all relevant URLs."
)
Step 2: Fetch Target Documentation
Use the path from Step 1, always with .md extension:
WebFetch(
url: "https://mlflow.org/docs/latest/[path].md",
prompt: "Return all code blocks verbatim. Do not summarize."
)
Anti-Patterns
Do not use .html files — Fetch .md source files only.
Do not use WebSearch — Always start from llms.txt; web search returns outdated or third-party content.
Do not use vague prompts — "Extract complete documentation" allows summarization. Use "Return all code blocks verbatim. Do not summarize."
Do not use versioned paths — Always use /docs/latest/, never /docs/3.8/ or other versions unless the user explicitly requests a specific version.
Do not guess URLs — Always verify paths exist in llms.txt before fetching. Never construct documentation paths from assumptions.
Do not follow external links — Stay within mlflow.org/docs. Do not follow links to GitHub, PyPI, or third-party sites.
Do not mix sources — Use only MLflow docs. Do not combine with LangChain docs, OpenAI docs, or other external documentation.
Do not use llms.txt for non-GenAI topics — The llms.txt index covers LLM/GenAI documentation only. For classic ML tracking features, paths may differ.
More from mlflow/skills
agent-evaluation
Use this when you need to EVALUATE OR IMPROVE or OPTIMIZE an existing LLM agent's output quality - including improving tool selection accuracy, answer quality, reducing costs, or fixing issues where the agent gives wrong/incomplete responses. Evaluates agents systematically using MLflow evaluation with datasets, scorers, and tracing. IMPORTANT - Always also load the instrumenting-with-mlflow-tracing skill before starting any work. Covers end-to-end evaluation workflow or individual components (tracing setup, dataset creation, scorer definition, evaluation execution).
251instrumenting-with-mlflow-tracing
Instruments Python and TypeScript code with MLflow Tracing for observability. Must be loaded when setting up tracing as part of any workflow including agent evaluation. Triggers on adding tracing, instrumenting agents/LLM apps, getting started with MLflow tracing, tracing specific frameworks (LangGraph, LangChain, OpenAI, DSPy, CrewAI, AutoGen), or when another skill references tracing setup. Examples - "How do I add tracing?", "Instrument my agent", "Trace my LangChain app", "Set up tracing for evaluation
231analyzing-mlflow-session
Analyzes an MLflow session — a sequence of traces from a multi-turn chat conversation or interaction. Use when the user asks to debug a chat conversation, review session or chat history, find where a multi-turn chat went wrong, or analyze patterns across turns. Triggers on "analyze this session", "what happened in this conversation", "debug session", "review chat history", "where did this chat go wrong", "session traces", "analyze chat", "debug this chat".
220mlflow-onboarding
Onboards users to MLflow by determining their use case (GenAI agents/apps or traditional ML/deep learning) and guiding them through relevant quickstart tutorials and initial integration. If an experiment ID is available, it should be supplied as input to help determine the use case. Use when the user asks to get started with MLflow, set up tracking, add observability, or integrate MLflow into their project. Triggers on "get started with MLflow", "set up MLflow", "onboard to MLflow", "add MLflow to my project", "how do I use MLflow".
219analyzing-mlflow-trace
Analyzes a single MLflow trace to answer a user query about it. Use when the user provides a trace ID and asks to debug, investigate, find issues, root-cause errors, understand behavior, or analyze quality. Triggers on "analyze this trace", "what went wrong with this trace", "debug trace", "investigate trace", "why did this trace fail", "root cause this trace".
217retrieving-mlflow-traces
Retrieves MLflow traces using CLI or Python API. Use when the user asks to get a trace by ID, find traces, filter traces by status/tags/metadata/execution time, query traces, or debug failed traces. Triggers on "get trace", "search traces", "find failed traces", "filter traces by", "traces slower than", "query MLflow traces".
217