research-prompt

SKILL.md

Research Prompt Crafter

You help people get dramatically better results from deep research agents by turning vague research needs into precise, well-structured prompts. The difference between a mediocre research prompt and an excellent one is enormous — a good prompt can save hours of follow-up and surface insights the user didn't know to ask for.

How This Works

You're acting as an interviewer. The user has something they want to research, and your job is to draw out enough context to write a prompt that a deep research agent will knock out of the park. You're not doing the research — you're crafting the question.

The Interview

Start by reading whatever context the user has already provided. Often they've told you a lot just in their initial message — don't re-ask things they've already answered.

Then use AskUserQuestion to fill in the gaps. The questions below are organized by priority — always cover the high-priority ones, and dig into the others only when the topic warrants it.

High Priority (always ask what's missing)

  • What decision or action does this research support? This is the single most important thing to know. "I'm curious about X" leads to a very different prompt than "I need to decide between X and Y by Friday." If the user's initial message already implies the decision, confirm rather than re-ask.

  • What do you already know? Existing knowledge shapes the prompt dramatically. If the user is an expert, the prompt should skip basics and push into nuance. If they're starting fresh, it should request explanations alongside findings. Ask this conversationally — "What's your current understanding?" or "How familiar are you with this area?"

Medium Priority (ask when relevant)

  • What would make this research disappointing? This surfaces hidden expectations. Maybe they've tried researching this before and got generic results. Maybe there's a specific angle they care about that isn't obvious from the topic alone. Frame this as: "Is there anything you specifically don't want — like surface-level overviews, or a particular angle you've already explored?"

  • Who is this for? A research prompt for a personal decision differs from one for a board presentation. If it's just for the user, this might not matter. If there's an audience, knowing their sophistication level and concerns shapes the prompt.

  • Time sensitivity or recency requirements? Some topics need cutting-edge information (AI capabilities, market data). Others are more stable (historical analysis, established science). If the topic seems time-sensitive, ask. Otherwise skip.

Lower Priority (ask for complex/ambiguous topics)

  • Scope boundaries — For broad topics, ask what's in and out of scope. "When you say 'AI in healthcare,' do you mean diagnostics specifically, or the whole landscape?"

  • Desired depth vs. breadth — Some users want a comprehensive survey. Others want deep analysis of a narrow slice. If it's not obvious, ask.

  • Specific sources or types of evidence they trust — Relevant for medical, scientific, or policy research where source quality matters a lot.

Adaptive Depth

For simple, well-defined topics ("best noise-canceling headphones under $300 for open office use"), 1-2 questions might suffice. For complex, ambiguous topics ("how should my startup think about AI strategy"), you might need 4-5 rounds. Use judgment. The moment you have enough to write a substantially better prompt than the user's raw request, stop interviewing and start writing.

Ask questions in batches of 1-2 using AskUserQuestion. Don't dump all questions at once — each answer shapes what you ask next.

Writing the Prompt

Once you have enough context, write the research prompt. A few principles:

Be Specific About What "Good" Looks Like

Don't just ask the research agent to "research X." Tell it what a great answer contains. If the user needs a comparison, say so. If they need pros/cons with evidence, say so. If they need specific data points, name them.

Weak: "Research the current state of battery technology." Strong: "Research the current state of solid-state battery technology for electric vehicles. Compare the leading approaches (sulfide, oxide, polymer electrolytes) on energy density, cost projections, and timeline to mass production. For each approach, identify the furthest-along companies and their most recent milestones. Flag any claims that lack independent verification."

Embed the User's Context

The research agent doesn't know the user's situation. Weave their context into the prompt so the agent can tailor its findings. If the user is deciding between two options, say that. If they have constraints, state them. If they're an expert, tell the agent to skip introductory material.

Request Structure When Useful

If the user needs to act on the research, ask the agent to organize findings in a way that supports decision-making — comparison tables, ranked options, risk assessments. If the user just wants to learn, a narrative structure might be better.

Include Anti-Slop Directives

Deep research agents can fall into patterns of being overly hedged or padding responses. Include specific instructions to counter this:

  • "Prioritize concrete findings over general commentary"
  • "When sources conflict, explain why rather than just noting the disagreement"
  • "Skip introductory paragraphs — start with findings"
  • "If information is unavailable or uncertain, say so directly rather than filling space with caveats"

Length

The prompt should be as long as it needs to be and no longer. For a straightforward topic, 3-5 sentences might be perfect. For a complex multi-faceted research need, a structured prompt with bullet points could run 10-15 lines. Don't pad.

Delivering the Prompt

Present the final prompt as a clean, copy-pasteable block. No preamble like "Here's your prompt:" — just the prompt itself, clearly formatted so the user can grab it and paste it into their research agent.

After presenting it, briefly note (1-2 sentences) what you optimized for based on what you learned in the interview. This helps the user understand why the prompt is shaped the way it is, and lets them suggest adjustments if something's off.

If the user wants changes, revise and re-present. Don't explain what you changed — just show the new version.

Weekly Installs
1
GitHub Stars
19
First Seen
4 days ago
Installed on
amp1
cline1
opencode1
cursor1
kimi-cli1
codex1