searching-mlflow-docs
MLflow Documentation Search
Workflow
- Fetch
https://mlflow.org/docs/latest/llms.txtto find relevant page paths - Fetch the
.mdfile at the identified path - Present results with verbatim code examples
Step 1: Fetch llms.txt Index
WebFetch(
url: "https://mlflow.org/docs/latest/llms.txt",
prompt: "Find links or references to [TOPIC]. List all relevant URLs."
)
Step 2: Fetch Target Documentation
Use the path from Step 1, always with .md extension:
WebFetch(
url: "https://mlflow.org/docs/latest/[path].md",
prompt: "Return all code blocks verbatim. Do not summarize."
)
Anti-Patterns
Do not use .html files — Fetch .md source files only.
Do not use WebSearch — Always start from llms.txt; web search returns outdated or third-party content.
Do not use vague prompts — "Extract complete documentation" allows summarization. Use "Return all code blocks verbatim. Do not summarize."
Do not use versioned paths — Always use /docs/latest/, never /docs/3.8/ or other versions unless the user explicitly requests a specific version.
Do not guess URLs — Always verify paths exist in llms.txt before fetching. Never construct documentation paths from assumptions.
Do not follow external links — Stay within mlflow.org/docs. Do not follow links to GitHub, PyPI, or third-party sites.
Do not mix sources — Use only MLflow docs. Do not combine with LangChain docs, OpenAI docs, or other external documentation.
Do not use llms.txt for non-GenAI topics — The llms.txt index covers LLM/GenAI documentation only. For classic ML tracking features, paths may differ.
More from b-step62/skills
agent-evaluation
Use this when you need to IMPROVE or OPTIMIZE an existing LLM agent's performance - including improving tool selection accuracy, answer quality, reducing costs, or fixing issues where the agent gives wrong/incomplete responses. Evaluates agents systematically using MLflow evaluation with datasets, scorers, and tracing. Covers end-to-end evaluation workflow or individual components (tracing setup, dataset creation, scorer definition, evaluation execution).
9querying-mlflow-metrics
Fetches aggregated trace metrics (token usage, latency, trace counts, quality evaluations) from MLflow tracking servers. Triggers on requests to show metrics, analyze token usage, view LLM costs, check usage trends, or query trace statistics.
8searching-mlflow-traces
Searches and filters MLflow traces using CLI or Python API. Use when the user asks to find traces, filter traces by status/tags/metadata/execution time, query traces, or debug failed traces. Triggers on "search traces", "find failed traces", "filter traces by", "traces slower than", "query MLflow traces".
7instrumenting-with-mlflow-tracing
Instruments code with MLflow Tracing for observability. Triggers on questions about adding tracing, instrumenting agents/LLM apps, getting started with MLflow tracing, or tracing specific frameworks (LangGraph, LangChain, OpenAI, DSPy, CrewAI, AutoGen). Examples - "How do I add tracing?", "How to instrument my agent?", "How to trace my LangChain app?", "Getting started with MLflow tracing
7