ml-experiment-tracking
SKILL.md
Ml Experiment Tracking
Overview
Use this skill to make ML experiments comparable, reproducible, and audit-friendly.
Scope Boundaries
- Use this skill when the task matches the trigger condition described in
description. - Do not use this skill when the primary task falls outside this skill's domain.
Shared References
- Reproducibility metadata rules:
references/reproducibility-metadata-rules.md
Templates And Assets
- Tracking schema template:
assets/experiment-tracking-schema-template.md
Inputs To Gather
- Required metadata fields (code/data/config/artifacts).
- Tooling constraints for run logging and artifact storage.
- Reproducibility requirements by project risk level.
- Comparison dimensions for model decisions.
Deliverables
- Experiment tracking schema and mandatory fields.
- Run comparison protocol.
- Reproducibility verification checklist.
Workflow
- Define required metadata with
assets/experiment-tracking-schema-template.md. - Validate sufficiency using
references/reproducibility-metadata-rules.md. - Enforce run logging and artifact lineage.
- Re-run selected experiments from metadata only.
- Publish reproducibility confidence and gaps.
Quality Standard
- Every decision-grade run is reproducible.
- Artifact lineage is complete and queryable.
- Comparison views are consistent across runs.
Failure Conditions
- Stop when runs cannot be reproduced from recorded metadata.
- Stop when artifact lineage is incomplete.
- Escalate when tracking gaps block release decisions.
Weekly Installs
4
Repository
kentoshimizu/sw…t-skillsGitHub Stars
4
First Seen
14 days ago
Security Audits
Installed on
gemini-cli4
opencode4
codebuddy4
github-copilot4
codex4
kimi-cli4