semantic-layer-setup

SKILL.md

Semantic Layer Setup Orchestrator

End-to-end workflow for building the Databricks semantic layer — Metric Views, Table-Valued Functions, and Genie Spaces — on top of a completed Gold layer.

Predecessor: gold-layer-setup skill (Gold tables must exist before using this skill)

Time Estimate: 3-4 hours for initial setup, 1-2 hours per additional domain

What You'll Create:

  1. Metric Views — YAML-based semantic definitions for each Gold table
  2. Table-Valued Functions (TVFs) — parameterized SQL functions for Genie
  3. Genie Spaces — configured with agent instructions, data assets, benchmark questions

File Organization

Artifact Output Path (from repo root)
Metric View YAML definitions src/{project}_semantic/metric_views/*.yaml
Metric View creation script src/{project}_semantic/create_metric_views.py
TVF SQL definitions src/{project}_semantic/table_valued_functions.sql
Genie Space JSON configs src/{project}_semantic/genie_configs/*.json
Genie deployment notebook src/{project}_semantic/deploy_genie_spaces.py
Combined Asset Bundle job resources/semantic/semantic_layer_job.yml
Bundle config additions databricks.yml (sync + resource references)

{project} = project name from Asset Bundle variables (e.g., wanderbricks).


Decision Tree

Question Action
Building semantic layer end-to-end? Use this skill — it orchestrates everything
Only need Metric Views? Read semantic-layer/01-metric-views-patterns/SKILL.md directly
Only need TVFs? Read semantic-layer/02-databricks-table-valued-functions/SKILL.md directly
Only need Genie Space setup? Read semantic-layer/03-genie-space-patterns/SKILL.md directly
Need Genie API automation? Read semantic-layer/04-genie-space-export-import-api/SKILL.md directly
Need to optimize Genie accuracy? Read semantic-layer/05-genie-optimization-orchestrator/SKILL.md directly

Mandatory Skill Dependencies

CRITICAL: Before generating ANY code for the semantic layer, you MUST read and follow the patterns in these common skills. Do NOT generate these patterns from memory.

Phase MUST Read Skill (use Read tool on SKILL.md) What It Provides
All phases common/databricks-expert-agent Core extraction principle: extract names from source, never hardcode
Metric Views common/databricks-python-imports Pure Python module patterns for helpers
Deployment common/databricks-asset-bundles Job YAML, deployment patterns
All phases common/naming-tagging-standards Dual-purpose COMMENTs, v3.0 TVF comments, enterprise naming
Troubleshooting common/databricks-autonomous-operations Deploy → Poll → Diagnose → Fix → Redeploy loop when jobs fail

Semantic-Domain Dependencies

Skill Requirement What It Provides
semantic-layer/01-metric-views-patterns MUST read at Phase 1 YAML syntax, validation, joins, window measures
semantic-layer/02-databricks-table-valued-functions MUST read at Phase 2 STRING params, Genie compatibility, null safety
semantic-layer/03-genie-space-patterns MUST read at Phase 3 7-section deliverable, agent instructions, benchmark Qs
semantic-layer/04-genie-space-export-import-api MUST read at Phase 3 (JSON config) and Phase 6 (API deployment) REST API JSON schema, programmatic deployment
semantic-layer/05-genie-optimization-orchestrator External — run separately after deployment Benchmark testing, 6 control levers, optimization loop

🔴 Non-Negotiable Defaults

Default Value Applied Where NEVER Do This Instead
Manifest required plans/manifests/semantic-layer-manifest.yaml Phase 0 — before any implementation ❌ NEVER create artifacts via self-discovery; STOP if manifest is missing
Metric View syntax WITH METRICS LANGUAGE YAML Every Metric View DDL ❌ NEVER use non-YAML metric views
TVF parameters All STRING type Every TVF signature ❌ NEVER use DATE, INT, or other non-STRING params (Genie incompatible)
Genie warehouse Serverless SQL Warehouse Every Genie Space ❌ NEVER use Classic or Pro warehouse
Benchmark questions Minimum 10 per Genie Space Every Genie Space ❌ NEVER deploy without benchmarks
Column comments Required on all Gold tables Before Genie Space creation ❌ NEVER create Genie Space without column comments

Working Memory Management & Progressive Disclosure

This orchestrator spans 7 phases (0–6). To maintain coherence without context pollution, follow these progressive disclosure principles from AgentSkills.io and Anthropic's context engineering guidance:

Just-in-Time Skill Loading (CRITICAL)

DO NOT read all worker skills at the start. Read each skill ONLY when you enter its phase:

  • Phase 1: Read 01-metric-views-patterns/SKILL.md → work → persist notes → discard skill from working memory
  • Phase 2: Read 02-databricks-table-valued-functions/SKILL.md → work → persist notes → discard
  • Phase 3: Read 03-genie-space-patterns/SKILL.md + 04-genie-space-export-import-api/SKILL.md → work → persist notes → discard
  • Phase 4-6: Read common/databricks-asset-bundles/SKILL.md → work → done

Each worker skill ends with a "Notes to Carry Forward" section that tells you exactly what to persist for downstream phases. Use those notes — not the full skill content — as your handoff.

Context Handoff Protocol

At each phase boundary, your working memory should contain ONLY:

  1. gold_inventory dict (from Phase 0 — persists through all phases)
  2. Previous phase's "Notes to Carry Forward" (structured summary of outputs)
  3. Current phase's worker skill (read just-in-time)

Discard after each phase: full YAML bodies, SQL source code, complete JSON configs — they are on disk and retrievable via file paths in the notes.

Phase Summary Notes

After each phase, persist a brief summary note capturing:

  • Phase 0 output: Manifest loaded, planning_mode, artifact counts, gold_inventory dict
  • Phase 1 output: Use "Metric Views Notes to Carry Forward" from 01-metric-views-patterns (MV names, paths, grain, measure counts, composability notes)
  • Phase 2 output: Use "TVF Notes to Carry Forward" from 02-databricks-table-valued-functions (TVF names, paths, parameter signatures, domain assignments)
  • Phase 3 output: Use "Genie Space Notes to Carry Forward" from 03-genie-space-patterns (space names, JSON paths, asset counts, benchmark counts)
  • Phase 4 output: Job YAML path, databricks.yml changes
  • Phase 5 output: Deployment status, job run ID, task statuses
  • Phase 6 output: API deployment status, space IDs for idempotent re-deployment

Why This Matters

Context is a finite resource with diminishing marginal returns. Each worker skill is 400-600 lines. Loading all 4 workers simultaneously (~2000 lines) would consume your attention budget on content irrelevant to the current phase. Progressive loading keeps each phase focused on the smallest set of high-signal tokens needed for that phase's work.


Phased Implementation Workflow

Phase 0: Read Plan — MANDATORY (5 minutes)

The semantic layer manifest is REQUIRED. Do NOT proceed without it.

This orchestrator implements exactly what the project plan defined — no more, no less. The manifest plans/manifests/semantic-layer-manifest.yaml is generated by the planning/00-project-planning skill (stage 5) and serves as the implementation contract.

🔴 If the manifest does not exist, STOP and tell the user:

"The semantic layer manifest (plans/manifests/semantic-layer-manifest.yaml) is missing. This orchestrator requires a project plan to define which Metric Views, TVFs, and Genie Spaces to create. Please run the planning/00-project-planning skill first (stage 5), then return here."

import yaml
from pathlib import Path

manifest_path = Path("plans/manifests/semantic-layer-manifest.yaml")

if not manifest_path.exists():
    raise FileNotFoundError(
        "REQUIRED: plans/manifests/semantic-layer-manifest.yaml not found. "
        "Run planning/00-project-planning (stage 5) first to generate the "
        "semantic layer manifest, then re-run this orchestrator."
    )

with open(manifest_path) as f:
    manifest = yaml.safe_load(f)

# Respect planning mode — workshop mode means strict artifact caps
planning_mode = manifest.get('planning_mode', 'acceleration')
if planning_mode == 'workshop':
    print("⚠️  Workshop mode active — creating ONLY the artifacts listed in the manifest")

# Extract implementation checklist from manifest
domains = manifest.get('domains', {})
total_mvs, total_tvfs, total_genie = 0, 0, 0
for domain_name, domain_config in domains.items():
    mvs = domain_config.get('metric_views', [])
    tvfs = domain_config.get('tvfs', [])
    genie = domain_config.get('genie_spaces', [])
    total_mvs += len(mvs)
    total_tvfs += len(tvfs)
    total_genie += len(genie)
    print(f"Domain {domain_name}: {len(mvs)} MVs, {len(tvfs)} TVFs, {len(genie)} Genie Spaces")

print(f"\nTotal: {len(domains)} domains, {total_mvs} Metric Views, "
      f"{total_tvfs} TVFs, {total_genie} Genie Spaces")

# Validate summary counts match actual artifact counts
summary = manifest.get('summary', {})
assert total_mvs == int(summary.get('total_metric_views', total_mvs)), \
    f"MV count mismatch: {total_mvs} actual vs {summary.get('total_metric_views')} in summary"
assert total_tvfs == int(summary.get('total_tvfs', total_tvfs)), \
    f"TVF count mismatch: {total_tvfs} actual vs {summary.get('total_tvfs')} in summary"
assert total_genie == int(summary.get('total_genie_spaces', total_genie)), \
    f"Genie count mismatch: {total_genie} actual vs {summary.get('total_genie_spaces')} in summary"

What the manifest provides:

  • domains{} — one entry per agent domain, each containing:
    • metric_views[] — name, source table, dimensions, measures, business questions
    • tvfs[] — name, parameters (all STRING), Gold tables used, business questions
    • genie_spaces[] — name, warehouse type, asset assignments, benchmark questions
  • summary — expected artifact counts for validation
  • planning_modeacceleration (full) or workshop (capped artifacts; do NOT expand via self-discovery)

Key principle: Create ONLY the artifacts listed in the manifest. Do NOT add Metric Views, TVFs, or Genie Spaces beyond what the plan specified. If the plan missed something, update the plan first — then re-run this orchestrator.

Gold Schema Extraction (Anti-Hallucination — MANDATORY)

After the manifest check, build a verified gold_inventory dict before any artifact creation begins. This dict is the ONLY source of table/column names for Phases 1-3. No artifact may reference a table or column not in this inventory.

Two-level extraction (defense in depth):

  1. Parse Gold YAML files from gold_layer_design/yaml/{domain}/*.yaml — extract table_name, columns[].name, columns[].type, primary_key, foreign_keys
  2. Query live catalogSELECT table_name, column_name, full_data_type FROM {catalog}.information_schema.columns WHERE table_schema = '{gold_schema}'
  3. Cross-reference YAML vs catalog — flag any discrepancies (tables in YAML not deployed, columns missing, type mismatches)

Build the gold_inventory dict:

gold_inventory = {
    "dim_customer": {
        "columns": {"customer_key": "BIGINT", "customer_name": "STRING", ...},
        "primary_key": ["customer_key"],
        "foreign_keys": []
    },
    "fact_sales": {
        "columns": {"sales_key": "BIGINT", "customer_key": "BIGINT", ...},
        "primary_key": ["sales_key"],
        "foreign_keys": [{"columns": ["customer_key"], "references": "dim_customer"}]
    }
}

Gate: The gold_inventory dict MUST be non-empty and cross-referenced before proceeding to Phase 1. If the catalog query returns zero tables, STOP and verify Gold tables are deployed.


Phase 1: Metric Views (1-2 hours)

Context setup: Read the skills below just-in-time. After this phase, persist the "Metric Views Notes to Carry Forward" and discard the full skill content.

# Skill Path What It Provides
1 data_product_accelerator/skills/common/databricks-expert-agent/SKILL.md Extract-don't-generate principle
2 data_product_accelerator/skills/common/naming-tagging-standards/SKILL.md CM-02 dual-purpose COMMENT format for Metric Views
3 data_product_accelerator/skills/common/databricks-python-imports/SKILL.md sys.path setup for creation script in Asset Bundle
4 data_product_accelerator/skills/semantic-layer/01-metric-views-patterns/SKILL.md YAML syntax, validation, joins, composability, format types

Input: For each domain in manifest['domains'], iterate over domain['metric_views']. Each entry defines name, source_table, dimensions, measures, and business_questions. Do NOT create Metric Views not listed in the manifest.

Steps:

  1. Read manifest['domains'][domain]['metric_views'] — this is your complete list of Metric Views per domain
  2. For each entry, use the manifest's source_table, dimensions, and measures — cross-reference every column against gold_inventory (Phase 0)
  3. Validation gate: For each YAML, apply ALL three validations:
    • Column existence: Verify every dimensions[].column and measures[].column exists in gold_inventory[source_table]["columns"]. Fail with explicit error listing unresolved references.
    • Transitive join detection: Inspect all join on clauses. If ANY join's on references a join alias instead of source, flag as transitive join error. Fix: restructure as nested joins (snowflake schema, DBR 17.1+) or use denormalized columns.
    • Format type validation: Verify all measure format.type values are one of: byte, currency, date, date_time, number, percentage. Common mistakes: percent (use percentage), decimal (use number).
  4. Create create_metric_views.py with sys.path setup from databricks-python-imports
  5. Test each Metric View with sample queries
  6. Track completion: check off each manifest entry as its Metric View is confirmed created

Phase 2: Table-Valued Functions (1-2 hours)

Context setup: Discard Phase 1 skill content. Keep only gold_inventory + Phase 1's "Metric Views Notes to Carry Forward". Read the skills below just-in-time.

# Skill Path What It Provides
1 data_product_accelerator/skills/common/databricks-expert-agent/SKILL.md Extract TVF names/columns from gold_inventory
2 data_product_accelerator/skills/common/naming-tagging-standards/SKILL.md CM-04 v3.0 structured TVF COMMENTs
3 data_product_accelerator/skills/semantic-layer/02-databricks-table-valued-functions/SKILL.md STRING params, null safety, Genie compat, notebook_task deployment

Input: For each domain in manifest['domains'], iterate over domain['tvfs']. Each entry defines name, description, parameters (all STRING), gold_tables_used, and business_questions. Do NOT create TVFs not listed in the manifest.

Steps:

  1. Read manifest['domains'][domain]['tvfs'] — this is your complete list of TVFs per domain
  2. For each entry, use the manifest's parameters (ALL STRING) and gold_tables_used — cross-reference against gold_inventory
  3. Implement TVFs with null safety and SCD2 handling
  4. Add v3.0 bullet-point comments per naming-tagging-standards CM-04
  5. Validation gate: Parse each TVF SQL to extract all table/column references. Verify every reference exists in gold_inventory. Fail with explicit error listing hallucinated references.
  6. Validate with test queries
  7. Track completion: check off each manifest entry as its TVF is confirmed created

Phase 3: Genie Space Setup (1 hour)

Context setup: Discard Phase 2 skill content. Keep gold_inventory + Phase 1 notes (MV names/paths) + Phase 2's "TVF Notes to Carry Forward" (TVF names/paths). Read the skills below just-in-time. Phase 3 uses TWO worker skills — Genie Space Patterns (for design) and Export/Import API (for JSON config generation).

# Skill Path What It Provides
1 data_product_accelerator/skills/common/databricks-expert-agent/SKILL.md Extract asset references from gold_inventory
2 data_product_accelerator/skills/common/naming-tagging-standards/SKILL.md Table/column COMMENTs required by Genie
3 data_product_accelerator/skills/semantic-layer/03-genie-space-patterns/SKILL.md 7-section deliverable, agent instructions, benchmark questions
4 data_product_accelerator/skills/semantic-layer/04-genie-space-export-import-api/SKILL.md JSON schema, array sorting, ID generation, idempotent deployment

Input: For each domain in manifest['domains'], iterate over domain['genie_spaces']. Each entry defines name, warehouse, assets (metric_views, tvfs, tables), benchmark_questions_count, and benchmark_questions. Do NOT create Genie Spaces not listed in the manifest.

Context from prior phases: Use Phase 1's MV notes to assign metric views to spaces. Use Phase 2's TVF notes to assign TVFs. Use gold_inventory for Gold table assignments.

Steps:

  1. Verify all Gold tables have column comments (Genie depends on these)
  2. Read manifest['domains'][domain]['genie_spaces'] — use the manifest's assets to assign data assets (Metric Views, TVFs, Gold Tables) to each space
  3. Write General Instructions (≤20 lines)
  4. Create benchmark questions — use the manifest's benchmark_questions as the baseline; ensure minimum 10 per space with exact SQL answers
  5. Validation gate: Parse each benchmark question's expected SQL. Verify all table/column references exist in gold_inventory. Verify all TVF references match TVFs created in Phase 2. Verify all Metric View references match MVs created in Phase 1. Fail with explicit error listing hallucinated references.
  6. Configure Serverless SQL Warehouse (as specified in manifest's warehouse field)
  7. Generate API-compatible Genie Space JSON config file using the genie-space-export-import-api JSON schema. Save to src/{project}_semantic/genie_configs/. Use template variables (${catalog}, ${gold_schema}) for portability.
  8. Track completion: check off each manifest entry as its Genie Space config is confirmed generated

Phase 4: Asset Bundle Configuration (30 min)

Context setup: Discard Phase 3 skill content. Keep gold_inventory + Phase 3's "Genie Space Notes to Carry Forward" (space names, JSON paths, space IDs). Read just-in-time:

# Skill Path What It Provides
1 data_product_accelerator/skills/common/databricks-asset-bundles/SKILL.md Job YAML patterns, serverless config, notebook_task vs sql_task, base_parameters

Activities:

  1. Copy semantic-layer-job-template.yml from data_product_accelerator/skills/semantic-layer/00-semantic-layer-setup/assets/templates/ to resources/semantic/semantic_layer_job.yml — customize paths and variables
  2. Add YAML/JSON sync to databricks.yml:
    sync:
      include:
        - "src/semantic/metric_views/**/*.yaml"
        - "src/semantic/genie_configs/**/*.json"
    
  3. Add resource reference to databricks.yml: resources/semantic/semantic_layer_job.yml
  4. Ensure warehouse_id variable in databricks.yml:
    variables:
      warehouse_id:
        description: "SQL Warehouse ID for notebook_task execution"
        default: ""
    
    The warehouse ID is required for notebook tasks that execute SQL. Retrieve it from the Databricks workspace SQL Warehouse settings page.
  5. Add per-Genie-Space ID variables for update-or-create pattern:
    variables:
      genie_space_id_<space_name>:
        description: "Existing Genie Space ID for <space_name> (empty for initial creation)"
        default: ""
    
    These variables enable idempotent deployments: if a space ID is provided, the deploy script PATCHes the existing space instead of creating a duplicate. After first deployment, record the space IDs and set them in variables.

Combined Job Structure (3 tasks with depends_on chains):

  • create_metric_viewsnotebook_task, no deps
  • create_table_valued_functionsnotebook_task, depends_on: create_metric_views
  • deploy_genie_spacesnotebook_task, depends_on: create_metric_views + create_table_valued_functions

⚠️ All 3 tasks use notebook_task. sql_task.parameters are SQL bind parameters (:param) — they cannot substitute identifiers in DDL like ${catalog}.${gold_schema} in CREATE FUNCTION statements.

Critical Rules (from databricks-asset-bundles):

  • notebook_task for Metric Views, TVFs, and Genie (all 3 tasks)
  • base_parameters dict for all notebook_task entries (variable substitution for catalog, schema, etc.)
  • warehouse_id required for tasks that execute SQL (pass via base_parameters to notebooks)
  • environment_version: "4" with PyYAML + requests dependencies
  • YAML/JSON sync is CRITICAL — without it, creation scripts cannot find configs

Output: resources/semantic/semantic_layer_job.yml, updated databricks.yml

See assets/templates/semantic-layer-job-template.yml for the starter template.

Phase 5: Deploy & Run (30 min)

Context setup: Keep Phase 4's job YAML path + all accumulated notes. Read just-in-time:

# Skill Path What It Provides
1 data_product_accelerator/skills/common/databricks-autonomous-operations/SKILL.md Deploy → Poll → Diagnose → Fix → Redeploy loop

Two commands — platform-enforced ordering:

  1. databricks bundle deploy -t dev
  2. databricks bundle run semantic_layer_job -t dev

Databricks enforces the depends_on chain: Metric Views are created first, then TVFs, then Genie Spaces. If any task fails, downstream tasks do not run.

Verification:

  • Check all 3 task statuses in the job run output
  • Verify Metric Views: SHOW VIEWS IN {catalog}.{gold_schema}
  • Verify TVFs: SHOW FUNCTIONS IN {catalog}.{gold_schema}
  • Verify Genie Spaces: check Genie UI or use export_genie_space.py --list

On failure: Follow the databricks-autonomous-operations diagnose → fix → redeploy loop.

Phase 6: API Deployment (Recommended, 30 min)

Context setup: If Phase 3's "Genie API Notes to Carry Forward" are still available, use them. Otherwise re-read just-in-time:

# Skill Path What It Provides
1 data_product_accelerator/skills/semantic-layer/04-genie-space-export-import-api/SKILL.md REST API, JSON schema, idempotent deployment, array sorting

Steps:

  1. Export Genie Space config from dev as JSON (or use the Phase 3 JSON)
  2. Parameterize with variable substitution (${catalog}, ${gold_schema})
  3. Import to staging/prod environment via REST API

This complements the Asset Bundle approach. Phase 5 deploys within a single workspace; Phase 6 enables cross-workspace promotion via the REST API.

Genie Space Optimization (Separate Step)

Genie Space optimization is performed separately after deployment. Use semantic-layer/05-genie-optimization-orchestrator/SKILL.md directly after the semantic layer deployment checkpoint has passed. This ensures the Genie Space is live and queryable before running benchmark tests.


Post-Creation Validation

Manifest Compliance (CRITICAL)

  • plans/manifests/semantic-layer-manifest.yaml was read at Phase 0 before any implementation
  • Every Metric View maps 1:1 to a domains[domain].metric_views[] entry in the manifest
  • Every TVF maps 1:1 to a domains[domain].tvfs[] entry in the manifest
  • Every Genie Space maps 1:1 to a domains[domain].genie_spaces[] entry in the manifest
  • No artifacts were created via self-discovery (only manifest-driven)
  • If planning_mode: workshop, artifact counts do NOT exceed manifest totals
  • Manifest summary counts match actual deployed artifact counts

Anti-Hallucination Compliance

  • gold_inventory dict built from YAML + catalog in Phase 0
  • All table/column references in Metric Views validated against gold_inventory
  • All table/column references in TVFs validated against gold_inventory
  • All benchmark SQL references validated against gold_inventory + Phase 1/2 outputs

Common Skill Compliance

  • Names extracted from gold_inventory (not generated) per databricks-expert-agent
  • All COMMENTs follow naming-tagging-standards dual-purpose format (CM-02)
  • TVF COMMENTs follow v3.0 structured format (CM-04)
  • Asset Bundle YAML follows databricks-asset-bundles patterns
  • Python imports follow databricks-python-imports sys.path setup

Semantic Layer Specifics

  • All Metric Views use WITH METRICS LANGUAGE YAML syntax
  • All TVFs use STRING parameters only
  • Genie Space has ≤20 line General Instructions
  • Genie Space has ≥10 benchmark questions with exact SQL
  • Genie Space uses Serverless SQL Warehouse
  • All Gold tables have column comments before Genie Space creation
  • API-compatible Genie Space JSON generated in src/{project}_semantic/genie_configs/

Deployment Compliance

  • Combined semantic_layer_job.yml with depends_on chains created
  • databricks.yml updated with sync (YAML + JSON) and resource references
  • All 3 tasks pass in semantic_layer_job run

Pipeline Progression

Previous stage: planning/00-project-planning → Project plan for semantic layer, observability, ML, and GenAI agent phases should be complete

Next stage: After completing the semantic layer, proceed to:

  • monitoring/00-observability-setup — Set up Lakehouse Monitoring, AI/BI Dashboards, and SQL Alerts

Related Skills

Skill Relationship Path
metric-views-patterns Mandatory — Metric View YAML semantic-layer/01-metric-views-patterns/SKILL.md
databricks-table-valued-functions Mandatory — TVF patterns semantic-layer/02-databricks-table-valued-functions/SKILL.md
genie-space-patterns Mandatory — Genie Space setup semantic-layer/03-genie-space-patterns/SKILL.md
genie-space-export-import-api Mandatory — JSON config + API deployment semantic-layer/04-genie-space-export-import-api/SKILL.md
genie-optimization-orchestrator External — Run separately after deployment semantic-layer/05-genie-optimization-orchestrator/SKILL.md
databricks-expert-agent Mandatory — Extraction principle common/databricks-expert-agent/SKILL.md
databricks-asset-bundles Mandatory — Deployment common/databricks-asset-bundles/SKILL.md
databricks-python-imports Mandatory — Python patterns common/databricks-python-imports/SKILL.md
naming-tagging-standards Mandatory — COMMENTs, naming, tags common/naming-tagging-standards/SKILL.md
databricks-autonomous-operations Mandatory — Deploy/diagnose/fix loop common/databricks-autonomous-operations/SKILL.md

Post-Completion: Skill Usage Summary (MANDATORY)

After completing all phases of this orchestrator, output a Skill Usage Summary reflecting what you ACTUALLY did — not a pre-written summary.

What to Include

  1. Every skill SKILL.md or references/ file you read (via the Read tool), in the order you read them
  2. Which phase you were in when you read it
  3. Whether it was a Worker, Common, Cross-domain, or Reference file
  4. A one-line description of what you specifically used it for in this session

Format

# Phase Skill / Reference Read Type What It Was Used For
1 Phase N path/to/SKILL.md Worker / Common / Cross-domain / Reference One-line description

Summary Footer

End with:

  • Totals: X worker skills, Y common skills, Z reference files read across N phases
  • Skipped: List any skills from the dependency table above that you did NOT need to read, and why (e.g., "phase not applicable", "user skipped", "no issues encountered")
  • Unplanned: List any skills you read that were NOT listed in the dependency table (e.g., for troubleshooting, edge cases, or user-requested detours)

Version History

  • v1.2.0 (Feb 2026) — Progressive disclosure protocol: just-in-time skill loading, context handoff between phases, explicit working memory management per AgentSkills.io and Anthropic context engineering best practices
  • v1.1.0 (Feb 2026) — TVF task type corrected to notebook_task; warehouse_id and Genie Space ID variables added; validation gates enhanced with transitive join detection and format type validation

References

Weekly Installs
1
GitHub Stars
2
First Seen
9 days ago
Installed on
amp1
cline1
opencode1
cursor1
kimi-cli1
codex1