handoff

Installation
SKILL.md

Handoff Prompt Generator

Generate a prompt that another agent can execute without guessing.

Choose the handoff mode

  • Use a shared-workspace handoff when the receiving agent can access the same repo, files, and artifacts.
  • Use a fresh-context handoff when the receiving agent starts cold, in another session, or on another platform.
  • Ask for the target model family if it is not implied by the user's request. If it still is not known, draft a vendor-neutral base prompt and mark any missing model-specific adjustments.

Read one model reference

Read only the reference that matches the receiving model:

Target model family Reference
OpenAI GPT-5.x / Codex references/openai.md
Anthropic Claude 4.x references/anthropic.md
Google Gemini 3.x references/google.md
xAI Grok 4.x / Grok Code references/xai.md

If the requested model version is newer than the reference, verify the latest official docs before drafting the handoff.

Gather only execution-critical context

Collect the minimum information that removes ambiguity:

  • objective
  • success criteria
  • scope boundaries
  • relevant files, commands, URLs, or artifacts
  • current state and known blockers
  • verification steps
  • output location or return format
  • coordination notes for parallel work

Do not pad the handoff with background that does not change the receiver's next action.

Build the base handoff

Use flat labeled sections. Prefer direct operational language over narrative explanation.

Shared-Workspace Handoff

Target model: [family/version]
Handoff type: shared-workspace sub-task

Objective
[One concrete outcome]

Success criteria
- [Observable completion condition]
- [Verification condition]

Context
- [Only facts needed for this slice of work]

Inputs and artifacts
- [file paths, branches, logs, docs, prior outputs]

Ownership
- [files or directories to modify]
- [areas to avoid]

Constraints
- [technical limits]
- [things the agent must not do]

Verification
- [commands, tests, or review checks to run]

Output
- [exact return format]
- [where to write or save artifacts]

Coordination
- [how this work fits with parallel tasks]

Fresh-Context Handoff

Target model: [family/version]
Handoff type: fresh context

Project
- name: [project name]
- overview: [1-2 sentences]
- entry points: [first files or docs to read]

Current state
- completed: [what is already done]
- remaining: [what still needs to be done]
- blockers/baseline: [known failures, risks, or assumptions]

Task
- objective: [single outcome]
- success criteria:
  - [observable condition]
  - [verification condition]

Constraints
- [scope limits]
- [things not to change]
- [environment or policy constraints]

Verification
- [commands, tests, or manual checks]

Output
- [exact deliverable shape]
- [how to report open questions or TODOs]

Use placeholders like [TODO: exact path] instead of inventing repository facts.

Apply model-specific tuning

After drafting the base handoff:

  • add only the adjustments from the matching reference file
  • prefer external runtime settings when the receiving harness exposes them
  • avoid inventing API-only controls inside plain chat prompts
  • generate one prompt per target model if the user wants multiple versions

Hold the quality bar

  • Keep the task atomic.
  • Define what "done" means.
  • Name files and commands whenever possible.
  • Reference shared artifacts by path instead of pasting large logs.
  • State explicit stop rules for destructive or broad changes.
  • Ask for findings first for review tasks.
  • Require source boundaries and citation expectations for research tasks.

Return format

When the user asks for a handoff prompt:

  1. Return the ready-to-send prompt in a fenced code block.
  2. List assumptions or placeholders after the prompt.
  3. Generate separate prompts when the user wants handoffs for multiple models.
Related skills
Installs
1
GitHub Stars
15
First Seen
Apr 1, 2026