spice-catalogs
Spice Catalog Connectors
Catalog connectors expose external data catalogs for federated SQL queries, preserving the source schema hierarchy.
Basic Configuration
catalogs:
- from: <connector>
name: <catalog_name>
params:
# connector-specific parameters
include:
- 'schema.*' # optional: filter tables
Supported Catalogs
| Connector | From Format | Description |
|---|---|---|
unity_catalog |
unity_catalog |
Databricks Unity Catalog |
databricks |
databricks |
Databricks with Spark Connect |
iceberg |
iceberg |
Apache Iceberg catalogs |
spice.ai |
spice.ai |
Spice.ai Cloud Platform |
glue |
glue |
AWS Glue Data Catalog |
Examples
Unity Catalog
catalogs:
- from: unity_catalog
name: unity
params:
unity_catalog_endpoint: https://my-workspace.cloud.databricks.com
databricks_token: ${ secrets:DATABRICKS_TOKEN }
include:
- 'my_schema.*'
Iceberg with S3
catalogs:
- from: iceberg
name: iceberg_catalog
params:
iceberg_catalog_type: rest
iceberg_rest_uri: https://my-iceberg-catalog.com
Spice.ai Platform
catalogs:
- from: spice.ai
name: spiceai
include:
- 'tpch.*'
Querying Catalog Tables
Tables are accessed using the full path: <catalog>.<schema>.<table>
SELECT * FROM unity.my_schema.customers LIMIT 10;
Include Patterns
Filter which tables to expose using glob patterns:
include:
- 'prod_schema.*' # all tables in prod_schema
- '*.customers' # customers table in any schema
- 'analytics.sales_*' # tables starting with sales_ in analytics
Documentation
More from spiceai/skills
spice-data-connector
Configure individual data source connectors in Spice — PostgreSQL, MySQL, S3, Databricks, Snowflake, DuckDB, GitHub, Kafka, and 25+ more. Use this skill whenever the user wants to add a dataset, connect to a specific database or data source, load data from S3 or files, configure connector-specific parameters, understand file formats (Parquet, CSV, PDF, DOCX), or set up hive partitioning. This skill is the reference for the `from:` and `params:` fields in dataset configuration. For cross-source federation, views, and catalogs, see spice-connect-data.
22spicepod-config
Create and configure Spicepod manifests (spicepod.yaml) — the central configuration file for Spice applications. Use this skill whenever the user wants to create a new spicepod.yaml from scratch, understand the overall spicepod structure and available sections, configure runtime settings (ports, caching, telemetry/observability), set up a complete Spice application combining datasets + models + search, or understand deployment models and use cases. This is the "glue" skill that shows how all Spice components fit together in one manifest. For details on specific sections (datasets, models, search, etc.), see the dedicated skills.
16spice-models
Configure AI/LLM model providers and connections in Spice — OpenAI, Anthropic, Azure, Google, xAI, Bedrock, Perplexity, Databricks, HuggingFace, and local GGUF models. Use this skill whenever the user wants to add a model, configure a specific LLM provider, set up an OpenAI-compatible endpoint (e.g. Groq, Ollama), serve a local model, configure system prompts, set parameter overrides (temperature, response format), or understand which providers are available. This skill is the model connector reference. For AI features like tools, memory, workers, and NSQL, see spice-ai.
16spice-accelerators
Choose and configure the right acceleration engine — Arrow, DuckDB, SQLite, Cayenne, PostgreSQL, or Turso. Use this skill whenever the user needs to pick an accelerator engine, compare engines (e.g. "should I use DuckDB or Cayenne?"), configure engine-specific parameters (duckdb_file, sqlite_file), tune memory vs file mode, or understand engine capabilities and limitations. This skill is the engine selection and tuning guide. For the broader acceleration feature (refresh modes, retention, snapshots, indexes), see spice-acceleration.
15spice-acceleration
Accelerate data locally for sub-second query performance — the feature and its configuration. Use this skill whenever the user asks about data acceleration concepts, enabling acceleration on a dataset, choosing refresh modes (full, append, changes, caching), configuring retention policies, setting up snapshots for cold-start, adding indexes and constraints, or understanding the difference between federated and accelerated queries. This skill covers the "what and why" of acceleration. For choosing which acceleration engine to use (Arrow vs DuckDB vs SQLite vs Cayenne), see spice-accelerators.
10spice-connect-data
Connect Spice to data sources and query across them with federated SQL — including datasets, catalogs, views, and writes. Use this skill whenever the user wants to set up federated queries across multiple sources, create views, configure catalogs (Unity Catalog, Databricks, Iceberg), write data with INSERT INTO, or understand how Spice's query federation works. This skill focuses on the federation layer — cross-source joins, views, catalogs, and data writes. For configuring individual data source connectors (PostgreSQL params, S3 file formats, etc.), see spice-data-connector.
9