extract

Installation
SKILL.md

stardust:extract

Crawl an existing website, parse each page, extract the brand surface, and produce a stardust-formatted snapshot of the current state under stardust/current/. The output describes what the site is; later sub-commands consume it to decide what it should be.

This skill is descriptive: it does not invent direction, it does not critique, and it does not modify the live site. It writes only under stardust/current/ and updates stardust/state.json.

Inputs

  • <url> — required. The origin to crawl. Examples: https://example.com, https://example.com/shop. A path narrows the same-origin crawl to that subtree.
  • --cap <N> — optional. Override the default 5-page cap. The cap is intentionally small — a 5-page sample (home + four IA pillars/templates) is enough for cross-page brand aggregation, system-component detection, and the brand-review HTML to do useful work. Lift the cap with --cap 25 (the previous default) or higher when a deeper crawl is genuinely needed.
  • --all — optional. Lift the cap entirely; extract every discovered page after junk filtering. Equivalent to --cap 0. Use when the user spontaneously asks for a full crawl.
  • --pages <slug,slug,...> — optional. Restrict the crawl to specific paths (slugs derived per reference/ia-extraction.md). Bypasses the cap.
  • --refresh <slug> — optional. Re-extract one page that already exists in state.json.
  • --single — optional. Equivalent to --cap 1. Useful for testing.
  • --wait <fast|medium|spec|auto> — optional. Wait strategy per page. Default medium. See reference/playwright-recipe.md § Wait modes.
  • --no-junk-filter — optional. Disable the default junk-page filter in discovery (see reference/ia-extraction.md § Filtering).
  • --no-consent-dismiss — optional. Skip the pre-flight consent / cookie banner dismissal (see reference/playwright-recipe.md § Pre-flight: consent dismissal). Use when the redesign scope explicitly includes the consent surface, or when the dismissal's side-effects (script activation that wouldn't otherwise run) need to be avoided. Default behaviour is to dismiss; the contract preserves screenshots, voice aggregation, and per-section style from being polluted by the banner.
  • --prep — optional. Run in migrate-prep mode: lift the cap, type each page, detect module candidates, capture typed content slots, emit the prep summary. See § Prep mode below. Typically invoked via the prepare-migration orchestrator skill rather than directly.

Setup

Run the master skill's setup procedure first (skills/stardust/SKILL.md § Setup): impeccable dep check, context loader, state read.

Additional checks for this sub-command:

  1. Playwright availability. The extraction step needs a real browser. Detect Playwright in this order: a Playwright MCP server, then npx playwright. If neither is available, stop and tell the user how to install Playwright.
  2. Origin collision. If stardust/state.json already records site.originUrl and the new <url> is a different origin, stop and ask before clobbering. Stardust does not silently mix two sites in one project.
  3. Browser context. Open a fresh BrowserContext for the run. Run the consent dismissal pre-flight per reference/playwright-recipe.md § Pre-flight: consent dismissal unless --no-consent-dismiss is set. Cookies persist across the per-page loop within the same context, so one dismissal covers the whole crawl. Record the resolved method in _crawl-log.json#consent.method.
  4. Bot-management probe. When the first navigation in the run returns ERR_HTTP2_PROTOCOL_ERROR or ERR_QUIC_PROTOCOL_ERROR, or hangs through the entire hard-cap on what should be a fast origin, do not retry headless. Switch to headless: false, channel: 'chrome' per reference/playwright-recipe.md § Bot-management fallback and record the switch in _crawl-log.json#discovery.fetchTechnique so re-runs start in headed mode without rediscovering the issue.

Procedure

Phase 1 — Discovery

Discover the page inventory before crawling. Procedure in reference/ia-extraction.md. In summary:

  1. Fetch <origin>/sitemap.xml, then <origin>/sitemap_index.xml, then check robots.txt for Sitemap: directives.

  2. If no sitemap is reachable, run a same-origin BFS crawl from <url>, depth-limited to 3, link-extracting from rendered HTML.

  3. Filter the discovered URL list: same origin only, exclude mailto:, tel:, anchor-only links, query-only variations, common asset paths (.css, .js, .pdf, image extensions).

  4. De-duplicate trailing-slash variations.

  5. Apply the junk-page filter (reference/ia-extraction.md § Junk-page filter) unless --no-junk-filter is set. Surface the filtered list to the user as overridable.

  6. Apply the cap (default 5, or --cap, or --all for no cap) and proceed silently. Print an informational summary of what was kept and what was cut — but do not gate on user confirmation. The default cap is small enough that the common case is "extract 5 pages and move on"; pausing for a yes/no reply on every run is friction without value. Users who want different scope set it spontaneously at command time:

    $stardust extract https://example.com              # default 5 pages
    $stardust extract https://example.com --cap 25     # bump to 25
    $stardust extract https://example.com --all        # lift the cap
    $stardust extract https://example.com --pages home,about,pricing
    $stardust extract https://example.com --single     # just the entry URL
    

    The agent reads spontaneous scope intent from the user's prompt (e.g. "extract all pages", "look at just the home and pricing", "do a full crawl") and applies the equivalent flag. No re-confirmation needed once intent is clear.

    Informational output (not a prompt — proceed immediately):

    Discovered 38 pages on https://example.com (sitemap.xml).
    Filtered as likely junk (5): /test/, /sample-page/, /holiday1/, ...
    Selecting 5 highest-priority pages:
      - / (home)
      - /about
      - /pricing
      - /products
      - /contact
    
    Cut (28 pages, --all to lift): /blog/post-1, /blog/post-2, ...
    
    Extracting...
    

    Selection heuristic: page-type checklist first, then score-based ranking (home + IA-pillar keywords + sitemap priority − archive / version markers). See reference/ia-extraction.md § Page selection and § Priority for the cap. The English-only keyword list is a known limitation for localized sites.

  7. Write the discovered list to stardust/current/_crawl-log.json (created if absent) with _provenance and the full discovery reasoning, including filteredAsJunk[] and userChoice. This is an audit trail, not a state file.

Phase 2 — Per-page extraction

For each page in the cap-respecting list, render with Playwright following reference/playwright-recipe.md. The recipe is mandatory — in particular, do not skip the wait, scroll, or capture-list steps:

  • Viewport 1440 × 900 @ 2× DPR
  • Wait per the configured wait mode (default medium; see § Wait modes in reference/playwright-recipe.md)
  • Disable animations via prefers-reduced-motion: reduce
  • After the wait resolves, scroll to bottom in 4 viewport-height steps with 300 ms pauses, then return to top — this is required to trigger lazy-load and IntersectionObserver-driven content
  • Record waitMs and waitMode in the per-page _provenance

Capture per page (full schema in reference/current-state-schema.md):

  • Page metadata (title, meta description, OG tags, theme-color)
  • Semantic structure: heading outline, landmark roles, sections
  • Content: visible text per section (full innerText, no truncation per reference/playwright-recipe.md § Capture list 7), structured paragraphs (body[]), lists, FAQ Q/A pairs, and review/testimonial quotes per § Capture list 7-bis. Without these structured fields, every body region under a heading falls back to placeholder signature at migrate time.
  • CTA labels and href targets, link inventory (internal vs external)
  • Per-section computed style summary: dominant colors, font families in use, spacing rhythm, border-radius, shadows
  • Media inventory: img/srcset with original URLs and intrinsic dimensions, inline SVG count, video/iframe presence, cssBackgrounds[] (including pseudo-element ::before/ ::after walks per § Capture list 11) so background-image heroes and motifs do not silently disappear from extract.
  • Font files captured via network-intercept (per § Capture list 16): every woff2/woff/ttf/otf response saved under assets/fonts/ and recorded in _brand-extraction.json#type.files[] with licensing flag.
  • Icon-font detection (per § Capture list 17): when the page uses [class^="icon-"] with non-default ::before font-family + codepoint, capture the family, save the file, and record the iconClass → codepoint table in _brand-extraction.json#iconFont.
  • Interactive elements: forms (with field types), buttons, modals detected by ARIA roles

Save to stardust/current/pages/<slug>.json with _provenance as the first key. Save referenced media to stardust/current/assets/media/ preserving basename plus a short content hash.

Live-render evidence (synthesis is forbidden). Refuse to mark a page extracted in state.json unless its _provenance contains renderedBy: "playwright", an ISO-8601 fetchedAt, a positive integer waitMs, a waitMode from the recipe, and a final httpStatus in the 2xx/3xx range. These five fields are the contract enforced by reference/current-state-schema.md § Live-render evidence and read back by every downstream phase via validateProvenance() per skills/stardust/reference/state-machine.md § Provenance validation. Synthesizing a page record from _brand-extraction.json plus URL patterns plus captured photos — the 2026-04-30 lovesac shortcut — is the failure mode this guard exists to prevent. When the agent (or a delegated sub- agent) cannot satisfy the contract for a page, treat the page as a Phase 2 failure: record under _crawl-log.json#crawl.failures[] with errorClass: "ProvenanceMissing" and continue.

Mark the page extracted in state.json immediately after each successful page write. If a page fails, record the error in _crawl-log.json and continue — extraction is best-effort per page.

Phase 3 — Brand-surface extraction

Run once, after Phase 2 has finished, so cross-page aggregation has data to work with. Produces stardust/current/_brand-extraction.json per reference/brand-surface.md. Some fields are home-only (logo, voice samples, register heuristic); the visual tokens that drive DESIGN.md (palette, radius, shadow, type) are aggregated across all extracted pages to avoid the home-page bias documented in brand-surface.md § Aggregation scope. Captures:

  • Logo by the v1 priority chain: inline SVG → <img> with logo-ish class/id → apple-touch-iconog:image → favicon → synthesized placeholder. Save to stardust/current/assets/logo.<ext>.
  • Palette — aggregate computed colors across all extracted pages (background, text, accents, borders, hovers). Frequency-sort, cluster near-duplicates, emit a role-named list (background, surface, text, primary, secondary, accent).
  • Type — font families in use with their weights, sizes, and computed line-heights. Identify the heading family vs body family. Run the modular-scale audit (brand-surface.md § Modular-scale audit) and emit scaleAudit.kind = "modular" | "ad-hoc".
  • Motifs — signature border-radius (cross-page mode of non-zero values, weighted by element count), shadow stack (top 3 distinct, cross-page), gradient inventory, common patterns (chip, badge, card, hero-with-image). When the home-only mode disagrees with the cross-page mode, surface the divergence in _provenance.notes.
  • Voice samples — first paragraph of body copy, the hero headline, 3 representative CTA labels, a representative link list. Used by direct later but extracted now so the network round-trip is over.
  • Hero image — elevate the home page's primary visual asset to voice.heroImage (per reference/brand-surface.md § heroImage resolution). Without this elevation, downstream prototype reasons over a 16-image list and frequently picks the og:image instead of the live hero.
  • Icon font — when detected per reference/playwright-recipe.md § Capture list 17, populate _brand-extraction.json#iconFont with family, file path, and the iconClass → codepoint table so prototypes can render the brand's actual icons.
  • System components — cross-page repeated DOM blocks (site header, site footer, cross-promo strips, persistent CTAs, breadcrumbs). Detected by heading-sequence + CTA-label fingerprint per reference/brand-surface.md § System components. Required — these are usually the most load-bearing surfaces and must not silently disappear from the redesign target.

Do not invent values. Every captured value cites a source selector or URL in _brand-extraction.json for traceability.

Phase 4 — Seed stardust/current/PRODUCT.md and DESIGN.md

The current-state PRODUCT.md and DESIGN.md are descriptive, not authored — there is no interview to run because the user is not defining intent here, the agent is describing the existing site. Write them directly using impeccable's format specs:

  • For PRODUCT.md, follow the section structure in impeccable's reference/teach.md. Populate Register from the brand surface (sites that read as marketing/landing → brand; tools/dashboards → product; ambiguous → brand with a note). Populate Users, Product Purpose, Brand Personality, Anti-references, and Design Principles from the captured copy and the brand surface. Where the agent must infer, mark the section with _provenance: inferred and a one-line basis sentence.
  • For DESIGN.md and DESIGN.json, follow the format spec in impeccable's reference/document.md. Populate frontmatter (colors, typography, rounded, spacing, components) from the captured tokens. The extensions block of DESIGN.json carries v1's componentStyle, motifs, and voice arrays so nothing is lost.

Stardust does not invoke $impeccable teach or $impeccable document for the current-state files: those commands write to project root (the target) and run an interview. Stardust authors the descriptive snapshot directly. The format spec from impeccable is the contract; the runtime command is not.

The target-state PRODUCT.md and DESIGN.md at the project root are written by $stardust direct in Phase 2 of the pipeline, not here.

Phase 5 — Render stardust/current/brand-review.html

After Phase 4 writes the descriptive PRODUCT.md and DESIGN.md, emit the current-state brand review per reference/brand-review-template.md.

The brand-review HTML is the first surface a human can eyeball to verify the extraction before committing to a redesign direction. Misreads in the JSON (a wrong dominant radius, a missing system component, a single-page palette bias) are obvious to the eye in five seconds and invisible in JSON until someone notices. Putting the review at the end of extract catches misreads while they are still cheap to fix — re-extract is fast; re-direct + re-prototype is not.

The template is mandatory. In particular:

  1. Run the Tensions detectors listed in reference/brand-review-template.md § Detectors. Each rule is mechanical; emit a tension card whenever the trigger condition matches. The review may ship with zero tensions if the data is too thin to evaluate, but the detectors must always be run.
  2. Render in the brand's own captured colors and fonts, not a stardust shell.
  3. Embed all CSS; do not load external JavaScript or fonts unless the live site already does.
  4. Cite the source artifact for every section (e.g. _brand-extraction.json § type under Typography).

If the data for a section is missing, omit the section — do not fabricate placeholders. The coverage callout at the top reflects what is missing.

Phase 6 — Update state and report

After all Phase 2-5 writes succeed:

  1. Update stardust/state.json (schema in skills/stardust/reference/state-machine.md):

    • site.originUrl, site.extractedAt, site.pageCap, site.totalDiscovered, site.crawled
    • pages[] — one entry per crawled page with status: "extracted", filled currentStatePath, empty prototypePath and migratedPath
  2. Print a one-screen summary:

    Extracted https://example.com (5/38 pages, sitemap.xml)
    
    stardust/current/
      PRODUCT.md            (register: brand, inferred from landing)
      DESIGN.md             (5 colors, 2 type families, 3 motifs)
      brand-review.html     (4 tensions surfaced)
      pages/                (5 files)
      assets/logo.svg       (extracted from inline SVG)
      _brand-extraction.json
      _crawl-log.json
    
    Per-page evidence:
      slug         live  waitMode               waitMs   status
      /            yes   medium                 2380     200
      /about       yes   medium                 2110     200
      /pricing     yes   medium                 1940     200
      /products    yes   medium                 2640     200
      /contact     yes   domcontentloaded(fb)   8000     200
    
    Wait summary: 4 resolved at medium (avg 2.4s), 1 fallback (timed out at 8s)
      → /contact may be under-captured; consider --refresh
    
    Open stardust/current/brand-review.html to verify the extraction
    before running $stardust direct.
    
    Coverage note: extracted 5 of 38 discovered pages. The brand
    surface and brand-review use cross-page aggregation, so 5 pages
    covering distinct templates is usually sufficient. To extract
    more, re-run with --cap <N> (e.g. --cap 25) or list specific
    slugs with --pages.
    
    Next: $stardust direct  (resolve a redesign direction)
    

    The per-page evidence table is mandatory. The live column is yes when _provenance.renderedBy === "playwright" AND waitMs > 0, else no. A no row means the page record was not produced by a live Playwright render — this should never happen given the write-time guard, but the visible column is the defense-in-depth signal that catches the failure mode when it does (the 2026-04-30 lovesac synthesis bug went four phases deep before being caught because no report column surfaced the missing provenance). A maintainer scanning the summary should see yes on every row.

    Compute the wait summary by grouping each page's _provenance.waitMode and averaging waitMs. List slugs whose waitMode ends in (fallback) (rendered as (fb) in the table for width) as candidates for --refresh.

Outputs

Path Purpose
stardust/current/PRODUCT.md Descriptive strategy of the existing site (impeccable format)
stardust/current/DESIGN.md Descriptive visual system (Stitch format)
stardust/current/DESIGN.json Sidecar with extensions for motifs, voice, components
stardust/current/brand-review.html Self-contained visual review of the extraction (first eyeball-able artifact)
stardust/current/pages/<slug>.json Per-page parsed structure + content
stardust/current/assets/logo.<ext> Extracted logo
stardust/current/assets/media/ Extracted media referenced by pages
stardust/current/assets/screenshots/ Per-page viewport screenshots (used by brand-review)
stardust/current/_brand-extraction.json Consolidated brand surface (palette, type, motifs, voice, system components)
stardust/current/_crawl-log.json Discovery + crawl audit trail
stardust/state.json Updated with site + per-page status

Concurrency

Per state-machine.md: stardust does not lock. Two concurrent extracts on the same project are last-write-wins. Document this in the user report; do not engineer around it.

Failure modes

  • Network failure mid-crawl. Continue, record in _crawl-log.json, end with a partial state. State.json reflects only successfully extracted pages. User can re-run; already-extracted pages are skipped unless --refresh <slug>.
  • HTTP 4xx/5xx, non-HTML content, soft-404s. Validated explicitly per reference/playwright-recipe.md § Response validation. Each produces a distinct error class (HTTPError, ContentTypeError, EmptyPageError) recorded in _crawl-log.json#crawl.failures[]. Failed pages do not appear in state.json as extracted — they appear only in the failure log. Without this validation a 5xx page silently lands as an empty success and propagates wrong data to direct and prototype.
  • Login wall. Do not attempt to authenticate. If the home page redirects to a login screen, capture that one page, mark the rest as unreachable, and ask the user how to proceed (provide cookies via Playwright config, change the entry URL, or scope to public pages).
  • Bot-management block (Akamai / Cloudflare / F5 / Imperva). When the first navigation returns ERR_HTTP2_PROTOCOL_ERROR, ERR_QUIC_PROTOCOL_ERROR, or hangs through the hard-cap on a TLS/H2 fingerprint check, the issue is JA3/H2 fingerprinting on bundled-chromium-default headless mode — not auth, not network. Switch to headless: false, channel: 'chrome' per reference/playwright-recipe.md § Bot-management fallback. Do not retry headless: it will fail identically. The headed fallback works against most enterprise / commerce origins; playwright-extra + stealth plugin is a non-standard escape hatch for the residual cases. The headed window pops visibly, which is acceptable for interactive runs and unacceptable for unattended pipelines — surface this to the user when first triggered.
  • JavaScript-only content. Playwright already handles this. If the configured wait condition never fires within the mode's hard cap (reference/playwright-recipe.md § Wait modes), fall back to domcontentloaded and capture what is rendered. Record the fallback in the per-page _provenance.waitMode and surface in the wait-summary line of the final report.
  • Synthesis attempt (forbidden). When the agent (or a delegated sub-agent) cannot run a real Playwright render for a page — whether due to time pressure, token budget, or a tool/network failure — the only correct outcome is to record the page as a Phase 2 failure (errorClass: "ProvenanceMissing" in _crawl-log.json#crawl.failures[]) and continue. Synthesizing a page record from _brand-extraction.json plus URL patterns plus captured photos at "semantically matching" template positions is forbidden. This was the 2026-04-30 lovesac.com failure — 20 of 25 pages synthesized this way and the cascade ran four phases on the synthesized data before the gap was caught by a meta-question. The synthesis shortcut produces output indistinguishable from a successful run and propagates fabricated content through every downstream phase.

Prep mode (--prep)

When invoked with --prep, extract runs an extended pass that prepares the inventory for migration. Discovery-mode runs (without --prep) are unchanged: small cap, no typing, no module detection, presales-friendly. --prep is the gesture that says "the user is committing to migrate; build the data structure migrate consumes."

--prep adds five things on top of the standard procedure:

1. Lift the cap

--prep implies --all. Migration coverage requires the full inventory — the small discovery cap (5 pages) is insufficient. The cap-respecting selection logic from reference/ia-extraction.md § Page selection still applies for ordering and junk-filtering; it just doesn't truncate.

Sub-agent prompt requirements (when delegating)

When --prep is heavy enough that the agent delegates extraction to a sub-agent (a presales-shaped pattern when the inventory is large), the sub-agent prompt must:

  1. Forbid synthesis by name. The literal sentence *"do not synthesize a page record from _brand-extraction.json
    • URL patterns + captured photos; every page must be a live Playwright render"* must appear in the prompt. The earlier wording "must actually invoke Playwright per page" was satisfiable in spirit by synthesis-with-photo-reuse and produced the lovesac failure. Naming the shortcut explicitly closes that loophole.
  2. Require a per-page evidence table in the return. Columns: slug | waitMode | waitMs | fetchedAt | httpStatus. The parent agent reads this table on completion and aborts if any row is missing or shows waitMs: 0.
  3. Require the wait-summary line in the return, formatted identically to Phase 6's wait summary, so the parent can surface it in the user-facing report without reformatting.

These three are mandatory; missing any of them in the sub-agent prompt is itself a recipe violation. The cascade-level guard in prepare-migration validates the resulting per-page JSONs via validateProvenance() regardless — but a well-formed sub-agent return makes the failure cheaper to diagnose.

2. Page typing

For each extracted page, infer the type field from URL pattern and content shape (LLM judgment). Catalog from skills/stardust/reference/state-machine.md § Page types: landing | article | listing | program | form | static | unique.

Write the inferred type to state.json.pages[].type. The user confirms or refines during direct --prep. Discovery-mode runs leave type as null.

3. Module candidate detection

After Phase 3 (brand-surface extraction), scan extracted pages for recurring structural patterns. A pattern that appears in N+ pages with similar shape (same sequence of elements, same data-section / data-purpose, similar text shape) is surfaced as a module candidate.

Signal-source priority

Detection consumes per-page captured fields in this priority order. Each higher signal is weighted more heavily in the match-score; lower signals are tie-breakers and corroboration, not primary evidence. The priority exists because higher-up fields are explicitly extracted and structured (no parsing ambiguity), while the bottom of the list (landmarks[].innerText substring search) is fragile against capture variations and was the source of the 2026-04-29 sliccy.com under-detection (0 hits for pre-footer-shell, 1 of 2 hits for install-tile — both modules genuinely present on every page, both invisible because the substrings being searched lived past the truncation boundary that has since been removed).

  1. pages/<slug>.json#headings[] — cross-page repeats of the same heading text in the same level. Highest signal: structured, explicit, captured in full regardless of body length.
  2. pages/<slug>.json#ctas[] labels — cross-page repeats of the same CTA label appearing on similar surfaces.
  3. pages/<slug>.json#media.cssBackgrounds[] URLs — same asset URL on multiple pages is a strong system-component signal (already specced as a system-component candidate in reference/brand-surface.md § Cross-page CSS-background reuse; module detection consumes the same signal at finer granularity).
  4. pages/<slug>.json#forms[] actions — cross-page repeats of the same form action URL. Newsletter / contact / search forms are the typical hits.
  5. pages/<slug>.json#components.componentsByLandmark when present (per future current-state-schema.md extension): per-landmark counts of cards / grids / etc.
  6. Substring search in landmarks[].innerText — lowest signal. Use only as corroboration once a candidate has already passed the higher-signal checks; never as the primary detector.

A candidate that fires on signals 1 + 2 above the threshold is high-confidence; a candidate that fires only on 6 should be treated as speculative and surfaced as such for the user to confirm in direct --prep.

Candidate output is a draft entry under DESIGN.json.extensions.modules[]:

{
  "id": "candidate-<short-hash>",
  "slots": [
    { "name": "<inferred>", "type": "text|link|image|...", "required": false }
  ],
  "instances": [
    { "slug": "home",   "selector": "..." },
    { "slug": "donate", "selector": "..." }
  ],
  "status": "candidate"
}

The status: "candidate" flag distinguishes draft entries from confirmed modules. direct --prep is where the user names them and promotes (or prunes).

4. Typed content slots

Per-page JSON (current/pages/<slug>.json) gains a slots section that identifies content slots per page-type:

  • article pages: headline, deck, byline, meta, lead-image, body, pullquotes[], related[]
  • listing pages: index-headline, filter-controls, card-grid with typed sub-slots per card
  • program pages: program-headline, summary, feature-grid, cta-band
  • landing, form, static — typed slots inferred per content shape

Schema additions live in reference/current-state-schema.md § Typed slots (extend that doc separately).

5. Prep summary

Replace Phase 6's standard report with the prep summary format:

extract --prep complete
=======================

Inventory:    127 pages crawled (5 prior, 122 new)
Provenance:   127/127 live (every page has Playwright evidence)
Page types:   landing 1 · article 84 · listing 6 · program 12 · form 3 · static 18 · unique 3
              (LLM-inferred; refine in direct --prep)

Module candidates: 8
  hotline-211         5 instances  (home, get-help, donate, news, programs)
  donate-band         12 instances (home, donate, news, all article footers)
  story-card          7 instances  (home, news, programs)
  ...

Typed slots:  filled per page-type (see current/pages/<slug>.json § slots)

Next: $stardust direct --prep  (confirm types, name modules)

The Provenance: <live>/<total> live line is mandatory in prep-mode output. When the ratio is anything other than <total>/<total> the prep run has failed the synthesis guard; list the affected slugs as a sub-bullet and treat the prep run as incomplete (the cascade-level guard in prepare-migration SKILL.md surfaces the same check between phases).

Default mode (no --prep) is unchanged. The flag is intended for the prepare-migration orchestrator, though direct invocation is supported.

References

  • reference/playwright-recipe.md — viewport, capture list, logo locator chain.
  • reference/ia-extraction.md — sitemap + BFS crawl + cap procedure.
  • reference/current-state-schema.md — per-page JSON schema.
  • reference/brand-surface.md — consolidated brand-surface schema.
  • reference/brand-review-template.md — current-state brand-review HTML contract + Tensions detectors.
  • skills/stardust/reference/state-machine.md — state.json contract.
  • skills/stardust/reference/artifact-map.md — provenance shape.
Related skills
Installs
28
Repository
adobe/skills
GitHub Stars
72
First Seen
8 days ago