data-pipeline-quality

SKILL.md

Testing Pyramid for Data

Run tests in this order. Cheapest and fastest first:

Layer What It Catches Examples
Schema tests (run first) Structural failures Column types, not-null, uniqueness, accepted values
Business rule tests Logic errors Cross-field validation, referential integrity, range checks
Integration tests (run last) System-level drift Cross-system reconciliation, end-to-end row counts

Schema tests are cheap. Run them on every pipeline execution. Business rule tests are mid-tier — run them on staging and production. Integration tests are expensive — run them on a schedule (daily or pre-release).

dbt Test Patterns

Generic tests for reusable checks. Apply across models:

models:
  - name: fct_encounters
    columns:
      - name: encounter_id
        tests: [not_null, unique]
      - name: encounter_type
        tests:
          - accepted_values:
              values: ['inpatient', 'outpatient', 'emergency', 'observation']
      - name: patient_id
        tests:
          - relationships:
              to: ref('dim_patient')
              field: patient_id

Custom generic test for row count tolerance:

{% test row_count_within_tolerance(model, min_count, max_count) %}
select count(*) as row_count
from {{ model }}
having count(*) < {{ min_count }} or count(*) > {{ max_count }}
{% endtest %}

Singular tests for business logic specific to one model. Use singular tests when the logic doesn't generalize.

Data Contracts

A data contract is a product spec for your data. It defines what consumers can depend on.

contract:
  name: fct_encounters
  version: 2
  owner: data-platform-team
  sla:
    freshness: "< 4 hours from source update"
    completeness: ">= 99.5% of expected rows"
    accuracy: ">= 99.9% match to source of record"
  schema:
    encounter_id: {type: bigint, nullable: false, unique: true}
    patient_id: {type: bigint, nullable: false}
    encounter_date: {type: date, nullable: false}

Producer responsibilities: Meet the SLA, notify consumers before breaking changes, version the schema.

Consumer expectations: Query only contracted fields, respect the grain, report quality issues.

Circuit Breakers

Wire quality gates into pipeline stages. When a check fails, block the pipeline and alert.

  • Schema violation: Block immediately. Bad schema corrupts everything downstream.
  • Row count outside tolerance: Block and alert. Investigate before proceeding.
  • Freshness SLA breach: Alert the on-call. Don't block unless downstream consumers can't tolerate stale data.
  • Business rule failure: Block if the failure rate exceeds threshold (e.g., >1% of rows). Alert if below.

Cross-reference data-quality-assessment for the 4-stage quality model (detect → assess → respond → prevent).

CRITICAL: Never auto-heal data quality issues in production. Alert, block, investigate. Auto-fixes mask root causes and erode trust faster than stale data.

Monitoring

Track quality over time, not just point-in-time pass/fail:

  • Row count trends: Expected vs actual with tolerance bands. A sudden 40% drop is a pipeline failure. A gradual 5% weekly decline is a source issue.
  • Freshness: Time since last successful pipeline run. Track P50 and P95, not just "last run."
  • Anomaly detection: Statistical thresholds beat static thresholds. A metric that normally ranges 1,000-1,200 firing at 950 is more useful than a static "alert below 500" that never triggers.

Cross-References

For quality scoring methodology (the 5-dimension rubric and maturity model), see data-quality-assessment. This skill covers how to automate those checks in your pipeline.

Weekly Installs
4
GitHub Stars
1
First Seen
Feb 21, 2026
Installed on
opencode4
gemini-cli4
github-copilot4
codex4
kimi-cli4
cursor4