together-embeddings
Together Embeddings & Reranking
Overview
Use this skill for semantic retrieval components:
- create embeddings
- batch embeddings
- build retrieval or RAG pipelines
- rerank retrieved candidates
This skill is for retrieval plumbing, not for the final language-model response itself.
When This Skill Wins
- Build vector search or semantic similarity features
- Add embedding generation to a data pipeline
- Improve retrieval quality with reranking
- Assemble a retrieval stage before calling a chat model
Hand Off To Another Skill
- Use
together-chat-completionsfor the final answer-generation step - Use
together-batch-inferencefor very large offline embedding backfills - Use
together-dedicated-endpointswhen reranking requires a dedicated deployment
Quick Routing
- Embeddings API usage
- Read references/api-reference.md
- Start with scripts/embed_and_rerank.py or scripts/embed_and_rerank.ts
- Semantic search (embed, store, query)
- Start with scripts/semantic_search.py -- includes an in-memory vector store, cosine-similarity retrieval, and optional rerank
- RAG pipeline composition
- Start with scripts/rag_pipeline.py
- Model selection and rerank constraints
- Read references/models.md
Workflow
- Confirm that the user needs vectors or retrieval, not direct generation.
- Choose the embedding model and batch shape.
- Generate embeddings for corpus and query paths consistently.
- Retrieve candidates. An in-memory cosine-similarity store works for prototyping and small corpora (see
semantic_search.py). Use a dedicated vector database for production scale. - Rerank only when the extra latency and endpoint requirement are justified. When no dedicated rerank endpoint is available, cosine-similarity ranking is a reasonable fallback.
High-Signal Rules
- Python scripts require the Together v2 SDK (
together>=2.0.0). If the user is on an older version, they must upgrade first:uv pip install --upgrade "together>=2.0.0". - Keep embeddings and reranking conceptually separate; rerank is a second-stage precision step.
- Reranking in this repo assumes a dedicated endpoint. Do not promise serverless rerank unless the product changes. When no endpoint is available, fall back to cosine-similarity ranking.
- The embedding model has a 514-token context limit. Chunk longer documents before embedding.
- The
rag_pipeline.pyexample demonstrates retrieval plus generation; treat generation as a hand-off to chat completions. - Preserve model consistency across indexing and querying.
Resource Map
- API details: references/api-reference.md
- Model guide: references/models.md
- Python embeddings example: scripts/embed_and_rerank.py
- TypeScript embeddings example: scripts/embed_and_rerank.ts
- Python semantic search: scripts/semantic_search.py
- Python RAG pipeline: scripts/rag_pipeline.py
Official Docs
More from zainhas/skills
together-audio
Use this skill for Together AI audio workflows: text-to-speech over REST, streaming, or realtime WebSocket APIs, plus speech-to-text transcription, translation, diarization, timestamps, and live transcription. Reach for it whenever the user needs audio in or audio out on Together AI rather than generic chat generation, image or video creation, or model training.
1together-images
Use this skill for Together AI image workflows: text-to-image generation, image editing with Kontext, FLUX model selection, LoRA-based styling, reference-image guidance, and local image downloads. Reach for it whenever the user wants to generate or edit images on Together AI rather than create videos or build text-only chat applications.
1together-video
Use this skill for Together AI video workflows: text-to-video generation, image-to-video with keyframe control, model and dimension selection, polling asynchronous jobs, and downloading completed videos. Reach for it whenever the user wants motion generation on Together AI rather than still-image generation or text-only inference.
1together-gpu-clusters
Use this skill for Together AI GPU clusters and raw infrastructure workflows: provisioning on-demand or reserved clusters, choosing Kubernetes vs Slurm, attaching shared storage, scaling, getting credentials, and operating cluster-backed ML or HPC jobs. Reach for it when the user needs multi-node compute or infrastructure control rather than a managed model endpoint.
1together-fine-tuning
Use this skill for Together AI fine-tuning workflows: LoRA or full fine-tuning, DPO preference tuning, VLM training, function-calling tuning, reasoning tuning, and BYOM uploads. Reach for it whenever the user wants to adapt a model on custom data rather than only run inference, evaluate outputs, or host an existing model.
1together-batch-inference
Use this skill for Together AI Batch API workflows: preparing JSONL inputs, uploading batch files, creating asynchronous jobs, polling status, downloading outputs, and optimizing large offline inference runs for lower cost. Reach for it whenever the user needs high-volume, non-interactive inference rather than real-time chat or evaluation jobs.
1