prompt-repetition
Prompt Repetition
When to use this skill
- The user is using a non-reasoning or lightweight model and accuracy drops when the important question appears late in the prompt.
- The prompt shape is long-context retrieval, options-first multiple choice, or position-sensitive lookup.
- The user wants a bounded decision rule for whether to duplicate the prompt 2× or 3×, not a vague promise that prompt engineering will help.
- The workflow needs a cheap experiment before changing models or building retrieval infrastructure.
- The request mentions things like the model forgot the question, missed the final instruction, lost the item at slot 25, got confused by a long options list, or performs worse on Haiku / Flash / mini models than on stronger reasoning models.
Do not use this skill as the main workflow when:
- The real problem is context selection, retrieval design, or noisy source stuffing → use broader context-engineering or RAG work.
- The real problem is reasoning depth and a reasoning-capable model is available.
- The prompt is tool-heavy / agentic / multi-step, where duplicating the full prompt may just multiply cost and confusion.
- The user needs a universal always-on middleware policy. This skill is for targeted use, not blanket auto-apply.
Core idea
Prompt repetition is a task-shape-specific intervention, not a universal prompting law.
The strongest evidence-backed cases are:
- Long-context retrieval — the question comes after a lot of context and the model misses late details.
- Options-first MCQ — choices appear before the question or the structure makes the actual ask easy to lose.
- Position-sensitive lookup — inventories, ordered lists, or record positions where the model loses track of index-like detail.
The skill succeeds when it answers three questions clearly:
- Is this one of the task shapes where repetition is worth testing?
- What repetition count is safe enough to try without blowing the context budget?
- What should we do instead if repetition is the wrong tool?
Instructions
Step 1: Classify the prompt shape before changing anything
Put the current prompt into one bucket:
| Shape | Typical signal | Repetition fit | Better alternative if fit is weak |
|---|---|---|---|
| Long-context retrieval | "The model ignores the question after a huge context block" | Good | Trim/select context, retrieval, restate question near end |
| Options-first MCQ | "The options come first and the model picks the wrong letter" | Good | Reorder question/options, simplify answer format |
| Position-sensitive lookup | "It misses item 25 in a long list" | Good | Chunk the list, provide structured table, use retrieval |
| Tool-heavy agent prompt | "There are many tools, policies, and steps" | Weak | Simplify policy, separate stages, use better orchestration |
| Reasoning-heavy task | "We need step-by-step synthesis or planning" | Weak | Use a reasoning model, decomposition, examples |
| RAG / search architecture | "The retrieved evidence is noisy or incomplete" | Weak | Fix retrieval, ranking, chunking, context packing |
If the task lands in a weak row, do not recommend prompt repetition as the main answer.
Step 2: Check the model and cost guardrail
Prompt repetition multiplies input tokens, even if the paper reports no increase in generated output length or latency for the tested runs.
Minimum checks:
- Model class — is this a non-reasoning / lightweight model or a stronger reasoning-capable model?
- Prompt length now — estimate whether 2× or 3× would still fit comfortably inside the usable context budget.
- Operational budget — is extra input cost acceptable for this workflow?
- Failure mode — are we fixing a late-context miss, or is the task actually about reasoning or retrieval quality?
Conservative default:
- if the prompt is already near the context limit, do not recommend full repetition
- if the task needs structured reasoning, do not use repetition as a substitute for model choice
- if only the final question/instruction matters, try restating the question before duplicating the whole prompt
Step 3: Choose the smallest useful intervention
Use the cheapest intervention that matches the failure:
-
Repeat the question only
- Best when the context is huge but only the final ask is being lost.
- Cheapest first experiment.
-
Repeat the full prompt 2×
- Default experiment for long-context retrieval and options-first MCQ on lightweight models.
- Good first pass when the whole prompt structure matters.
-
Repeat the full prompt 3×
- Reserve for clearly position-sensitive or index-heavy tasks.
- Only if the cost and context window are still safe.
If the user cannot afford the input-token overhead, route to prompt restructuring or retrieval instead.
Step 4: Apply explicit opt-out rules
Do not recommend repetition when any of these are true:
- the user already has a good reasoning model and the failure is multi-step reasoning
- the prompt includes many tools, policies, or action constraints that would just be duplicated noisily
- the retrieval set is weak, contradictory, or irrelevant
- the context budget is already tight enough that duplication risks truncation or degraded performance
- the task can be fixed more cleanly by moving the question, reducing context, or adding a small example
Step 5: Return a decision, not just a trick
A good answer should include:
- Fit — why repetition is or is not appropriate here
- Smallest recommended intervention — repeat question only, full prompt 2×, full prompt 3×, or do not use repetition
- Budget note — input-token / context-window implication
- Fallback / route-out — what to do instead if repetition is the wrong tool
Use this response shape:
"This is a [good / weak] fit for prompt repetition because [task shape]. I would try [smallest intervention] first. Watch [token/context risk]. If that fails, switch to [better alternative]."
Step 6: Keep evaluation narrow and comparable
When testing repetition:
- use the same prompt except for the repetition change
- compare baseline vs 2× vs 3× only when the budget allows
- judge success on task accuracy, not on style or length alone
- stop if the repeated prompt causes context pressure or obvious prompt bloat
If more than one change is needed, repetition is probably not the main lever.
Step 7: Route out when another job starts
- Context selection / retrieval / chunking / source quality → broader context-engineering or RAG work
- Reasoning-heavy synthesis / planning → stronger reasoning model or decomposition workflow
- Prompt examples / output-format steering → few-shot or prompt-structure work
- Tool-policy / action-constraint confusion → simplify orchestration instead of duplicating the whole prompt
Examples
Example 1: Options-first MCQ on a lightweight model
Prompt:
I have a long options-first multiple-choice prompt on a Flash model. Should I duplicate the prompt or restructure it?
Good response shape:
- classify as options-first MCQ
- recommend either question restatement or full prompt 2× as the first bounded experiment
- mention token-cost tradeoff
- note that reordering the prompt is a valid structural alternative
Example 2: RAG quality complaint
Prompt:
Our RAG answers are weak. Should we add prompt repetition everywhere?
Good response shape:
- say this is a weak fit
- explain that retrieval/context quality is the bigger lever
- route to context engineering or RAG fixes instead of blanket repetition
Example 3: Position-sensitive inventory lookup
Prompt:
A mini model keeps missing the item at slot 25 in a long inventory list.
Good response shape:
- classify as position-sensitive lookup
- allow 2× and possibly 3× repetition if budget permits
- mention structured-table or chunking fallback if cost is too high
Example 4: Reasoning model question
Prompt:
We have a reasoning-capable model available, but it is more expensive. Does prompt repetition replace it?
Good response shape:
- say no, repetition is not a substitute for reasoning ability
- limit repetition to targeted non-reasoning failure modes
- route multi-step reasoning to the stronger model or decomposition
Best practices
- Treat prompt repetition as a targeted experiment, not default middleware.
- Start with the smallest intervention that matches the failure mode.
- Always mention input-token cost and context-budget pressure.
- Prefer prompt restructuring or retrieval fixes when the real issue is context quality.
- Drop weakly evidenced claims instead of letting the skill become a generic prompt-hacks bucket.
- Keep the final recommendation binary: try repetition here or route elsewhere.