databricks-core-workflow-b

SKILL.md

Databricks Core Workflow B: MLflow Training

Overview

Build ML pipelines with MLflow experiment tracking, model registry, and deployment.

Prerequisites

  • Completed databricks-install-auth setup
  • Familiarity with databricks-core-workflow-a (data pipelines)
  • MLflow and scikit-learn installed
  • Unity Catalog for model registry (recommended)

Instructions

Step 1: Feature Engineering with Feature Store

Step 2: MLflow Experiment Tracking

Step 3: Model Registry and Versioning

Step 4: Model Serving and Inference

For full implementation details and code examples, load: references/implementation-guide.md

Output

  • Feature table in Unity Catalog
  • MLflow experiment with tracked runs
  • Registered model with versions
  • Model serving endpoint

Error Handling

Error Cause Solution
Model not found Wrong model name/version Verify in Model Registry
Feature mismatch Schema changed Retrain with updated features
Endpoint timeout Cold start Disable scale-to-zero for latency
Memory error Large batch Reduce batch size or increase cluster

Resources

Next Steps

For common errors, see databricks-common-errors.

Examples

Basic usage: Apply databricks core workflow b to a standard project setup with default configuration options.

Advanced scenario: Customize databricks core workflow b for production environments with multiple constraints and team-specific requirements.

Weekly Installs
18
GitHub Stars
1.6K
First Seen
Feb 14, 2026
Installed on
codex18
opencode17
github-copilot17
kimi-cli17
gemini-cli17
amp17