article-to-cover
article-to-cover
Overview
Art-director-level poster design skill. Analyzes text input, anchors style direction, plans visual hierarchy, and outputs structured AI generation instructions. Two modes: creative direction from scratch (no reference image), or reference-image-based reconstruction/mimicry.
Dependencies
- engine: meitu-ai
- user data: ~/.openclaw/visual/
Core Workflow
Step 1: Load Context
- Analyze user input — Read the provided text (article, chat transcript, or design brief). Synthesize core message into headline + subtitle candidates.
- User preferences — If
~/.openclaw/visual/memory/global.mdexists, read for style preferences. Ifcontexts/poster.mdexists, read for poster-specific preferences. - Experience log — If
~/.openclaw/visual/journal/knowledge.yamlexists, scan for entries related to this industry or similar tasks. - Brand assets — If user specifies a brand, read
~/.openclaw/visual/assets/brands/{brand}/:- Brand tone, personality, market positioning
- Color system (primary 60–70%, secondary 20–30%, accent 5–10%)
- Logo file and usage specs
- Derivative graphics if available
Step 2: Route — Determine Scenario
- User provides reference image → go to Step 3B: Poster Analyse
- No reference image (text/brief only) → go to Step 3A: Creative Direction
Step 3A: Creative Direction (No Reference Image)
Read references/design-constraints.md for hard rules (logo, human diversity, negative lexicon, medium-type). These apply to all output.
3A.1: Classify Brand Information Level
| Level | User provides | Action |
|---|---|---|
| B1 | Brand tone + color system | Anchor on tone and colors |
| B2 | Tone + colors + logo | Additionally analyze logo for visual linkage opportunities |
| Basic | No brand assets | Extract temperament from text, use industry mapping for color |
| Franchise | References a well-known IP (Harry Potter, Marvel, etc.) | Lock franchise visual DNA as style anchor, skip industry mapping |
3A.2: Identify Industry + Market
Industry identification — Read references/industry-styles.md for the mapping table:
- Match industry keywords from text content (food, fitness, finance, etc.)
- Validate via semantic analysis: core nouns, core verbs, usage scenarios
- If brand info exists, infer industry from brand attributes
- Output: primary industry + core semantic features (B2B/B2C, audience, emotional tone, scenario)
- If industry not in library → follow Unknown Industry Handling in same file
Market identification — Based on input language, cultural cues (currency, date format, festivals), brand origin:
- Output: target market (North America / Europe / Latin America / Asia Pacific) + cultural sensitivity requirements
3A.3: Determine Medium Type
Follow illustration trigger rules strictly from references/design-constraints.md:
- Illustration allowed ONLY if: user explicitly requests it, reference images are illustration-style, or user specifies illustration keywords
- Otherwise: must use Photography / Vector Graphic / 3D Rendering
- Apply negative lexicon: ban watercolor, line art, etching, hand-drawn terms
- Record medium type internally (do not show user)
3A.4: Anchor Style
- User provided brand logo → anchor on logo temperament, explore adjacent aesthetics within same visual school
- User specified style keywords → anchor on that style, explore adjacent range
- No style guidance → use industry auto-mapping from references/industry-styles.md, select highest-matching style
- For the 6 special industries (primary school, university, medical, nonprofit, finance, rental) → apply mandatory layout rules from industry-styles.md
Validate: visual clarity, emotional resonance, brand fit, execution feasibility, style–visual binding.
3A.5: Creative Ideation + Deepening
Read references/creative-framework.md for the full creative methodology. Execute in order:
- Deconstruct brief — Identify core assets and constraints (text, brand, imagery, layout)
- Style + element matching — Combine industry core subjects (coffee → machine/cup; beauty → products/brushes) with style signature elements to build visual scene
- Concept expansion — Avoid clichés; seek metaphorical visual expressions; balance metaphor with clarity; apply contrast, white space, rhythm, hierarchy, visual metaphor
- World-building — Typography as design protagonist or deep graphic interaction; scene construction from classic style scenarios
- If Franchise — Lock franchise visual DNA, label directions as "IP Name + Core Style – Variant", include franchise signature elements in core visual
- Deepen — If logo exists → deep graphic analysis (shape derivation: literalization / negative space / repetitive composition). Elaborate visual intention (composition + color + lighting → narrative + emotion). Apply Swiss International Style for layout system.
- Translate to AI instructions — Convert all artistic concepts to production instructions with style catalysts (era / medium texture / emotional aesthetics keywords)
3A.6: Output
Generate structured output following Scenario 1 format in references/output-formats.md:
- Design Direction with style name
- Core Visual (must include: style signature elements + industry core subject + stylized scene)
- Visual Elements (subject & environment, lighting & atmosphere, color language, composition & camera)
- Layout & Typography (typography concept, layout strategy with information hierarchy, text–image relationship)
- Overview (style + one-sentence summary of strongest visual scene)
- AI Production Instructions (JSON with project_manifest, visual_style_system, scene_elements, typography_layout, ai_generation_prompts)
Quality check before output:
- Output language matches user input language
- Strong binding between style name and all visual/layout/typography modules
- Core visual contains all three required elements
- Total colors ≤ 3 (excluding grayscale), body text contrast ≥ 4.5:1
Step 3B: Poster Analyse (With Reference Image)
Read references/design-constraints.md for hard rules. Read references/poster-analyse.md for the full analysis methodology.
3B.1: Intent Routing (Priority 1 > 2 > 3)
- Explicit commands — "like this image" / "keep layout" / "series" → Mimicry; "redesign" / "refer to vibe" / "optimize" → Washing
- Implicit scenarios — Content swap only ("replace person with cat", "change title") → Mimicry; Extract attribute for new carrier ("use this color scheme for something else") → Washing
- Ambiguity default — Reference image + simple keywords only → Washing (provide upgraded scheme, not copy-paste)
3B.2: Reverse-Engineer Reference Image
Extract comprehensive visual DNA:
- Style/medium — Physical texture only (e.g., "3D render", "Risograph"). Strictly forbidden to describe specific objects.
- Layout — Grid structure, composition logic, reading path
- Font form — Case (ALL CAPS / Title Case / lowercase), arrangement (stacked / curved / scattered)
- Brush stroke — Precise medium description (chalk texture, gouache dry brush, vector gradient), never generic "illustration"
- Detail insight — If no facial features → add "faceless character, blank face" to prompt and "eyes, nose, mouth" to negative
- Vector/stroke — Distinguish "flat vector, lineless, clean edges" vs "outlined, ink stroke"; if no strokes → emphasize "no outlines"
3B.3: Extract Soul Anchor (Washing mode only)
Identify the single most irreplaceable, highest-design-value element:
| Anchor type | Trigger | Action |
|---|---|---|
| Typography | Font is highly distinctive (liquid, 3D inflated) | Lock font style → reconstruct layout + color |
| Layout | Grid is highly distinctive (deconstructionism, special segmentation) | Lock layout framework → reconstruct font + color |
| Vibe | Light/shadow or medium is highly distinctive (film grain, acid light) | Lock physical texture → reconstruct layout + font |
3B.4: Deep Thinking
- Image-text layout — Brainstorm 6 options, select 1 distinctly different from reference
- Main visual — Use user description if provided; otherwise deduce from theme
- Text — Use user copy verbatim (modification strictly forbidden); if none, deduce from theme
- Color:
- Mimicry → follow reference colors
- Washing → execute Hue Cleansing: forbidden to reuse reference main hue; retain color relationship but replace hue (black+gold → white+chrome; high-sat red+blue → high-sat purple+green)
3B.5: Reconstruct
Mimicry mode: Style 100% locked. Redraw composition for new content. For text changes: identify original title A, confirm user's new copy B, output instruction "change text '[A]' to text '[B]'".
Washing mode:
- Coordinate Reset — Detect reference composition logic, select opposing logic:
- Symmetrical → negative space / diagonal / scattered
- Flat/front view → top-down / bottom-up / 3D / fisheye
- Real-scene photo → macro close-up / partial crop / out-of-focus
- Hue Cleansing — New colors must differ ≥ 90° on color wheel from reference
- Abandon reference grid completely. Build new reading path from user copy hierarchy.
3B.6: Self-Correction Protocol
Before outputting, verify:
- Color check — Does new palette overlap reference main hue? If yes + no user brand color specified → enforce color inversion or complementary color
- Layout check (Washing only) — Has layout been re-planned? If not → re-plan
- If any check fails → roll back and re-plan. Do not output failed result.
3B.7: Output
Generate JSON following Scenario 2 format in references/output-formats.md:
mission_logic— intent + soul extraction reasoningdesign_blueprint— reconstructed concept, style, color, layout, typography, composition, detected modecontent_firewall— discarded reference objects + unlocked featuresprompt— Final English prompt:[New Color]::3 + [concept]::2 + [anchor] + [mutations]. --no [ignored] --iw 0.5
Step 4: Generate Image via meitu-ai tool skill
- Extract prompt from Step 3A JSON (
ai_generation_prompts.primary_prompt) or Step 3B JSON (prompt) - Determine dimensions from design spec or user requirements (default: 1080×1350 portrait poster)
- Run:
python3 "{baseDir}/../meitu-ai/scripts/run_command.py" \ --command "image-generate" \ --input-json '{"prompt":"{prompt}","size":"{width}x{height}"}' - If generation fails → adjust prompt and retry through the same meitu-ai command runner
Step 5: Compliance Check
- Brand — Logo placement matches rules, colors accurate, tone consistent
- Platform — If target platform specified, read
~/.openclaw/visual/rules/platforms/{platform}.yamlfor size/format requirements - Content safety — Human diversity rules, no incomplete bodies, max 3 people in scene
- Readability — Body text contrast ≥ 4.5:1, total colors ≤ 3 (excluding grayscale)
Step 6: Record to Journal
- Append task entry to
~/.openclaw/visual/journal/entries/with: date, input summary, style direction chosen, output path, key decisions - Ask user: "要不要记录这次设计经验到知识库?" If yes → update
~/.openclaw/visual/journal/knowledge.yaml
Output
- Structured design specification (markdown + JSON)
- Generated poster image(s) via meitu-ai
- Journal entry (if user opts in)
More from tangyang/skills
meitu-image-edit
图片编辑能力。按提示词编辑已有图片,使用内置命令 image-edit。
30meitu-ai
meitu-ai 工具能力 Skill。集中维护安装、凭证、命令映射与统一执行入口。其他场景 Skill 按需调用本 Skill 的脚本。
19meitu-image-cutout
抠图能力。去除图片背景并输出前景结果,使用内置命令 image-cutout。
18meitu-image-generate
图片生成能力。根据提示词生成图片,可选参考图,使用内置命令 image-generate。
10meitu-image-virtual-tryon
试衣能力。将衣服图穿到人物图上,使用内置命令 image-virtual-tryon。
9meitu-image-upscale
图片超清能力。提升图片清晰度和分辨率,使用内置命令 image-upscale。
8