together-fine-tuning
Together Fine-Tuning
Overview
Use Together AI fine-tuning when the user needs to adapt a model to their own data or behavior.
Supported workflows in this repo:
- LoRA fine-tuning
- full fine-tuning
- DPO preference tuning
- VLM fine-tuning
- function-calling fine-tuning
- reasoning fine-tuning
- BYOM upload paths
When This Skill Wins
- Train a model on custom instruction or conversational data
- Improve function-calling reliability with supervised examples
- Train on preferences rather than only demonstrations
- Fine-tune multimodal or reasoning-oriented models
- Deploy a fine-tuned output model later through dedicated endpoints
Hand Off To Another Skill
- Use
together-chat-completionsfor plain inference without training - Use
together-evaluationsto measure a model before or after tuning - Use
together-dedicated-endpointsto host the resulting tuned model - Use
together-gpu-clustersonly when the user needs raw infrastructure rather than managed tuning
Quick Routing
- Standard LoRA or full fine-tuning
- Start with scripts/finetune_workflow.py
- Read references/data-formats.md
- DPO preference tuning
- Start with scripts/dpo_workflow.py
- Function-calling tuning
- Start with scripts/function_calling_finetune.py
- Reasoning tuning
- Start with scripts/reasoning_finetune.py
- VLM tuning
- Start with scripts/vlm_finetune.py
- Model support and deployment options
Workflow
- Choose the tuning method that matches the desired behavior change.
- Validate dataset format before spending tokens on training.
- Upload training data and keep the returned file ID.
- Create the job with explicit method-specific parameters.
- Monitor job state, events, and checkpoints before handing off to deployment.
High-Signal Rules
- Python scripts require the Together v2 SDK (
together>=2.0.0). If the user is on an older version, they must upgrade first:uv pip install --upgrade "together>=2.0.0". - Prefer LoRA unless the user has a specific reason to pay for full fine-tuning.
- Keep data-format validation close to the upload step so bad files fail early.
- Treat deployment as a separate phase; fine-tuning success does not automatically mean serving success.
- Use the method-specific script instead of overloading one generic workflow for all modes.
- Parameterize dataset paths, model IDs, and suffixes in automation instead of embedding one demo dataset forever.
Resource Map
- Data formats: references/data-formats.md
- Supported models: references/supported-models.md
- Deployment guide: references/deployment.md
- LoRA or full workflow: scripts/finetune_workflow.py
- DPO workflow: scripts/dpo_workflow.py
- Function-calling workflow: scripts/function_calling_finetune.py
- Reasoning workflow: scripts/reasoning_finetune.py
- VLM workflow: scripts/vlm_finetune.py
Official Docs
More from togethercomputer/skills
together-chat-completions
Real-time and streaming text generation via Together AI's OpenAI-compatible chat/completions API, including multi-turn conversations, tool and function calling, structured JSON outputs, and reasoning models. Reach for it whenever the user wants to build or debug text generation on Together AI, unless they specifically need batch jobs, embeddings, fine-tuning, dedicated endpoints, dedicated containers, or GPU clusters.
40together-evaluations
LLM-as-a-judge evaluation framework on Together AI. Classify, score, and compare model outputs, select judge models, use external-provider judges or targets, poll results and download reports. Reach for it whenever the user wants to benchmark outputs, grade responses, compare A/B variants, or operationalize automated evaluations.
31together-video
Text-to-video and image-to-video generation via Together AI, including keyframe control, model and dimension selection, asynchronous job polling, and video downloads. Reach for it whenever the user wants motion generation on Together AI rather than still-image generation or text-only inference.
29together-sandboxes
Remote Python execution in managed sandboxes on Together AI with stateful sessions, file uploads, data analysis, chart generation, and notebook-like runs via the Sandboxes API. Reach for it whenever the user wants managed remote Python execution instead of local execution, raw clusters, or full model hosting.
29together-dedicated-endpoints
Single-tenant GPU endpoints on Together AI with autoscaling and no rate limits. Deploy fine-tuned or uploaded models, size hardware, and manage endpoint lifecycle. Reach for it whenever the user needs predictable always-on hosting rather than serverless inference, custom containers, or raw clusters.
29