together-fine-tuning
Installation
SKILL.md
Together Fine-Tuning
Overview
Use Together AI fine-tuning when the user needs to adapt a model to their own data or behavior.
Supported workflows in this repo:
- LoRA fine-tuning
- full fine-tuning
- DPO preference tuning
- VLM fine-tuning
- function-calling fine-tuning
- reasoning fine-tuning
- BYOM upload paths
When This Skill Wins
- Train a model on custom instruction or conversational data
- Improve function-calling reliability with supervised examples
- Train on preferences rather than only demonstrations
- Fine-tune multimodal or reasoning-oriented models
- Deploy a fine-tuned output model later through dedicated endpoints
Hand Off To Another Skill
- Use
together-chat-completionsfor plain inference without training - Use
together-evaluationsto measure a model before or after tuning - Use
together-dedicated-endpointsto host the resulting tuned model - Use
together-gpu-clustersonly when the user needs raw infrastructure rather than managed tuning
Quick Routing
- Standard LoRA or full fine-tuning
- Start with scripts/finetune_workflow.py
- Read references/data-formats.md
- DPO preference tuning
- Start with scripts/dpo_workflow.py
- Function-calling tuning
- Start with scripts/function_calling_finetune.py
- Reasoning tuning
- Start with scripts/reasoning_finetune.py
- VLM tuning
- Start with scripts/vlm_finetune.py
- Model support and deployment options
Workflow
- Choose the tuning method that matches the desired behavior change.
- Validate dataset format before spending tokens on training.
- Upload training data and keep the returned file ID.
- Create the job with explicit method-specific parameters.
- Monitor job state, events, and checkpoints before handing off to deployment.
High-Signal Rules
- Python scripts require the Together v2 SDK (
together>=2.0.0). If the user is on an older version, they must upgrade first:uv pip install --upgrade "together>=2.0.0". - Prefer LoRA unless the user has a specific reason to pay for full fine-tuning.
- Keep data-format validation close to the upload step so bad files fail early.
- Treat deployment as a separate phase; fine-tuning success does not automatically mean serving success.
- Use the method-specific script instead of overloading one generic workflow for all modes.
- Parameterize dataset paths, model IDs, and suffixes in automation instead of embedding one demo dataset forever.
Resource Map
- Data formats: references/data-formats.md
- Supported models: references/supported-models.md
- Deployment guide: references/deployment.md
- LoRA or full workflow: scripts/finetune_workflow.py
- DPO workflow: scripts/dpo_workflow.py
- Function-calling workflow: scripts/function_calling_finetune.py
- Reasoning workflow: scripts/reasoning_finetune.py
- VLM workflow: scripts/vlm_finetune.py
Official Docs
Related skills
More from zainhas/togetherai-skills
together-images
Text-to-image generation and image editing via Together AI, including FLUX and Kontext models, LoRA-based styling, reference-image guidance, and local image downloads. Reach for it whenever the user wants to generate or edit images on Together AI rather than create videos or build text-only chat applications.
14together-sandboxes
Remote Python execution in managed sandboxes on Together AI with stateful sessions, file uploads, data analysis, chart generation, and notebook-like runs via the Sandboxes API. Reach for it whenever the user wants managed remote Python execution instead of local execution, raw clusters, or full model hosting.
1