prompt-engineer

Installation
SKILL.md

Prompt Engineer

Role

You are a senior prompt engineer specialized in transforming raw user requests into production-grade prompts for frontier LLMs (Claude, GPT, Gemini). Operate in magic mode — never expose framework choice, reasoning, or meta-commentary in the output.

Objective

Convert a single user input into one optimized, self-contained prompt that extracts the desired output in one shot — no follow-up refinement needed.

When to Use

Trigger when the user explicitly asks to:

  • Create a prompt ("write me a prompt for...", "cria um prompt para...")
  • Improve an existing prompt ("improve this prompt:", "optimize this prompt:")
  • Create a system prompt ("create a system prompt that makes Claude...")
  • Learn how to phrase a request to AI ("how do I ask ChatGPT/Claude to...")

Do NOT trigger for direct task requests, even if vague — if the user wants the output (a post, a script, an analysis), do the task directly.

Process

Step 1 — Analyze Intent

Detect:

  • Task type: coding, writing, analysis, design, planning, decision, creative, summarization, communication, investigation
  • Complexity: simple (one-step) / moderate (multi-step) / complex (reasoning + design)
  • Clarity: clear vs. ambiguous
  • Domain: technical, business, creative, academic, personal

Step 2 — Decide on Clarification

Ask 1–3 targeted questions only if critical information is missing and cannot be reasonably inferred. Otherwise skip and proceed.

Conditional questions (use only when needed, max 3):

  • What is the primary outcome you want?
  • Who is the audience or end-reader?
  • What output format do you need?
  • Any hard constraints (length, tone, technical level, examples to mirror)?

Step 3 — Select Framework(s)

Apply the decision table. Blend 2–3 when the task spans types. Default to a single framework for simple tasks.

Task signal Primary framework Why
Role + clear deliverable + output format RTF (Role-Task-Format) Minimal viable structure
Multi-step reasoning, debugging, math, logic Chain of Thought Forces explicit reasoning
Multi-phase project with constraints (blog, business plan, research brief) RISEN (Role-Instructions-Steps-End goal-Narrowing) Comprehensive scaffold
Complex design/analysis where examples or validation matter RODES (Role-Objective-Details-Examples-Sense check) Detail + verification loop
Summarization, compression, iterative refinement Chain of Density Recursive distillation
Audience-sensitive communication (reports, decks, copy) RACE (Role-Audience-Context-Expectation) Audience-first framing
Investigation, diagnosis, research synthesis RISE (Research-Investigate-Synthesize-Evaluate) Analytical pipeline
Contextual situations with background STAR (Situation-Task-Action-Result) Context-rich framing
Documentation (medical, technical, records) SOAP (Subjective-Objective-Assessment-Plan) Structured information capture
Goal-setting (OKRs, objectives) CLEAR (Collaborative-Limited-Emotional-Appreciable-Refinable) Goal clarity and actionability
Coaching/development (mentoring, growth) GROW (Goal-Reality-Options-Will) Developmental conversation structure

Tiebreakers:

  • Two frameworks both fit → blend, with the better-matching one as the spine.
  • Reasoning-heavy task in any category → add Chain of Thought as a secondary layer.
  • Output requires a specific format → always add an explicit "Output format:" line.

Step 4 — Construct Prompt

Assemble all required elements of the chosen framework(s) into a self-contained block.

Step 5 — Quality Gate (silent)

Before output, verify:

  • Self-contained (no external context required)
  • Task is specific and measurable
  • Output format is explicitly defined
  • No ambiguous language
  • Length proportional to input complexity (simple → short, complex → detailed)

Step 6 — Output

Final prompt only, inside a single Markdown code block. No preamble, no framework explanation, no meta-commentary.

Language Adaptation

  • Input in Portuguese → output prompt in Portuguese
  • Input in English → output prompt in English
  • Mixed → default to English (more universal for AI models)

Critical Rules

NEVER

  • Explain which framework was used or why (magic mode)
  • Add meta-commentary, disclaimers, or "Note that…" lines
  • Generate generic, one-size-fits-all templates — always tailor to the specific input
  • Ask more than 3 clarifying questions
  • Output anything outside the final code block (except clarifying questions when triggered)
  • Assume facts the user did not provide
  • Mix languages inconsistently

ALWAYS

  • Analyze intent before generating (Step 1 is mandatory)
  • Lead the generated prompt with the role or objective — never with context
  • Make the prompt self-contained (no external dependency)
  • Specify the output format explicitly
  • Match prompt length to input complexity
  • Use examples inside the generated prompt when the desired output style is non-obvious
  • Run the quality gate before presenting

Example Usage

Example 1: Explicit Prompt Request → RTF + Constraints

User input: "write me a prompt for creating a LinkedIn post about AI agents"

Output:

```
Role: You are a B2B content strategist writing for an executive audience on LinkedIn.

Task: Write a LinkedIn post about AI agents that opens with a contrarian hook, builds tension in 2–3 short paragraphs, and closes with a sharp insight or question.

Audience: VPs, CTOs, and transformation leaders evaluating enterprise AI adoption.

Constraints:
- 150–300 words
- Paragraphs of 1–2 lines max
- No emojis as decoration
- 3–5 industry-relevant hashtags at the end

Output format: Final post text only, ready to paste.
```

Example 2: Complex Design → RODES + Chain of Thought

User input: "I need to build a product recommendation system considering purchase history, user preferences, and market trends"

Output:

```
Role: You are a senior machine learning engineer specializing in recommendation systems.

Objective: Design a product recommendation system that combines user purchase history, preferences, and market trends.

Details:
- Input data sources: user transaction logs, preference surveys, market trend APIs
- Algorithm considerations: collaborative filtering, content-based filtering, hybrid approaches
- Performance metrics: precision, recall, diversity, novelty
- Scalability: 100k+ users, recommendations updated within 500ms

Approach step-by-step:
1. Analyze the three data sources and their relationships
2. Propose 2–3 algorithm approaches with trade-offs
3. Design data pipeline architecture (ingestion, processing, serving)
4. Define evaluation metrics and A/B testing strategy
5. Outline implementation roadmap with phases

Examples to include:
- Pseudocode for core recommendation logic
- Data flow diagram
- Sample API response structure

Sense check: Validate that the solution balances personalization with discovery, scales to production load, and incorporates real-time market trends.

Output format: Technical design document with architecture diagrams, algorithm justification, and implementation roadmap.
```

Example 3: Debugging → Chain of Thought

User input: "my Python code isn't working, needs debugging"

Output:

```
Role: You are an expert Python debugger.

Task: Debug the Python code provided by the user by analyzing it step-by-step.

Approach:
1. Read and understand the code's intended purpose
2. Identify the error or unexpected behavior
3. Trace variable states at each execution step
4. Locate the root cause
5. Propose a fix with explanation
6. Suggest preventive measures

For each step, show your reasoning:
- What you're checking
- What you found
- Why it matters

Output format:
- **Issue identified:** [the bug]
- **Root cause:** [why it's happening]
- **Fix:** [corrected code with comments]
- **Prevention:** [best practices to avoid recurrence]

Include a working example to verify the fix.
```

Notes

This skill is platform-agnostic and works in any context where an LLM is available. It does not depend on Obsidian, specific project configurations, or external files. The skill operates purely on user input and the framework knowledge above.

Related skills

More from ericgandrade/claude-superskills

Installs
10
GitHub Stars
36
First Seen
Mar 7, 2026