shinka-run
Shinka Run CLI Skill
Run a batch of program mutations using ShinkaEvolve's CLI interface.
When to Use
Use this skill when:
evaluate.pyandinitial.<ext>already exist- The user wants to run code evolution using the ShinkaEvolve/Shinka library
- You want configurable program evolution runs using explicit CLI args
Do not use this skill when:
- You need to scaffold a new task from scratch (use
shinka-setup)
What is ShinkaEvolve?
A framework developed by SakanaAI that combines LLMs with evolutionary algorithms to propose program mutations, that are then evaluated and archived. The goal is to optimize for performance and discover novel scientific insights.
Repo and documentation: https://github.com/SakanaAI/ShinkaEvolve Paper: https://arxiv.org/abs/2212.04180
Workflow
- Inspect task directory
ls -la <task_dir>
Confirm evaluate.py and initial.<ext> exist.
- Inspect CLI reference quickly
shinka_run --help
- Check model availability before proposing a run
shinka_models
shinka_models --verbose
Validate the exact run config against shinka_models:
- Mutation models: every entry in
evo.llm_modelsmust appear in thellmlist. - Meta recommendation models: if
evo.meta_rec_intervalis set andevo.meta_llm_modelsis set, every meta model must appear in thellmlist. - Prompt evolution models: if
evo.evolve_prompts=true, useevo.prompt_llm_modelswhen provided, otherwiseevo.llm_models; every selected model must appear in thellmlist. - Embedding model: if
evo.embedding_modelis set, it must appear in theembeddinglist. - Local OpenAI-compatible models are allowed for LLMs and embeddings via
local/<model>@http(s)://host[:port]/v1, and these local models are not expected to appear inshinka_models.
Important runtime rules:
- Do not assume meta recommendations fall back to
evo.llm_models. In the current runner, meta recommendations are only enabled whenevo.meta_llm_modelsis explicitly set. - Prompt evolution does fall back to
evo.llm_modelswhenevo.prompt_llm_modelsis unset. - Treat
local/<model>@http(s)://host[:port]/v1values as an explicit exception to theshinka_modelsmembership check. Instead, confirm the local endpoint URL and serving status separately before running. - If any required model is missing from
shinka_models, stop and ask the user to either change the config or set the missing credentials first.
- Confirm first-batch configuration with the user
- Minimum: budget scope, generation count, critical overrides.
- Explicitly confirm the mutation LLMs, meta recommendation LLMs, prompt evolution LLMs, and embedding model after checking them against
shinka_models. - If unclear, ask before running.
- Do not override any non-confirmed arguments.
- Launch main run with explicit knobs
shinka_run \
--task-dir <task_dir> \
--results_dir <results_dir> \
--num_generations 40 \
--set db.num_islands=3 \
--set job.time=00:10:00 \
--set evo.task_sys_msg='<task-specific system message guiding search>'\
--set evo.llm_models='["gpt-5-mini","gpt-5-nano"]' \
--set evo.meta_llm_models='["gpt-5-mini"]' \
--set evo.prompt_llm_models='["gpt-5-mini"]' \
--set evo.embedding_model='text-embedding-3-small' \
# Concurrency settings for parallel sampling and evaluation
--max-evaluation-jobs 2 \
--max-proposal-jobs 2 \
--max-db-workers 2
- Verify outputs before handoff
ls -la <results_dir>
Expect artifacts like run log, generation folders, and SQLite DBs.
- Between-batch handoff (unless explicitly autonomous)
- Summarize outcomes from the finished batch.
- Ask user for the next batch config before running again.
- Explicitly ask: "What new directions should we push next batch? Please include algorithm ideas, constraints, and failure modes to avoid."
- Turn user feedback into a revised system prompt and pass it via
--set evo.task_sys_msg=...in the nextshinka_runcall. - If the prompt is long/multiline, put it in a config file and use
--config-fnameinstead of shell-escaping. - Unless the user explicitly wants a fresh run/fork, keep the same
--results_dirfor follow-up batches.
Example next-batch command with feedback-driven prompt:
shinka_run \
--task-dir <task_dir> \
--results_dir <results_dir> \
--num_generations 20 \
--set evo.task_sys_msg='<new system prompt derived from user feedback>' \
--set db.num_islands=3
Batch Control Policy (Required)
Treat one shinka_run invocation as one batch of program evaluations/generations.
- Default mode: human-in-the-loop between batches.
- After each batch and before the first, always ask the user what configuration to run next (budget,
--num_generations, model/settings overrides, concurrency, islands, output path). - Do not start the next batch until the user confirms the next config.
- Keep
--results_dirfixed across continuation batches so Shinka can reload prior results. - Exception: if the user explicitly asks for fully autonomous execution, you may continue across batches without re-asking between runs.
More from sakanaai/shinkaevolve
shinka-setup
Create ShinkaEvolve task scaffolds from a target directory and task description, producing `evaluate.py` and `initial.<ext>` (multi-language). Use when asked to set up new ShinkaEvolve tasks, evaluation harnesses, or baseline programs for ShinkaEvolve.
102shinka-inspect
Load top-performing Shinka programs into agent context using `shinka.utils.load_programs_to_df`, and emit a compact Markdown bundle for iteration planning.
101shinka-convert
Convert an existing codebase in the current working directory into a ShinkaEvolve task directory by snapshotting the relevant code, adding evolve blocks, and generating `evaluate.py` plus Shinka runner/config files. Use when the user wants to optimize existing code with Shinka instead of creating a brand-new task from a natural-language description.
100