ux-for-ai
UX for AI — Design Framework
Use When
- AI interface design framework for building AI-powered features that feel premium, trustworthy, and world-class — not sloppy or robotic. Covers trust/transparency principles, avoiding AI slop, human oversight requirements, onboarding AI features...
- The task needs reusable judgment, domain constraints, or a proven workflow rather than ad hoc advice.
Do Not Use When
- The task is unrelated to
ux-for-aior would be better handled by a more specific companion skill. - The request only needs a trivial answer and none of this skill's constraints or references materially help.
Required Inputs
- Gather relevant project context, constraints, and the concrete problem to solve.
- Confirm the desired deliverable: design, code, review, migration plan, audit, or documentation.
Workflow
- Read this
SKILL.mdfirst, then load only the referenced deep-dive files that are necessary for the task. - Apply the ordered guidance, checklists, and decision rules in this skill instead of cherry-picking isolated snippets.
- Produce the deliverable with assumptions, risks, and follow-up work made explicit when they matter.
Quality Standards
- Keep outputs execution-oriented, concise, and aligned with the repository's baseline engineering standards.
- Preserve compatibility with existing project conventions unless the skill explicitly requires a stronger standard.
- Prefer deterministic, reviewable steps over vague advice or tool-specific magic.
Anti-Patterns
- Treating examples as copy-paste truth without checking fit, constraints, or failure modes.
- Loading every reference file by default instead of using progressive disclosure.
Outputs
- A concrete result that fits the task: implementation guidance, review findings, architecture decisions, templates, or generated artifacts.
- Clear assumptions, tradeoffs, or unresolved gaps when the task cannot be completed from available context alone.
- References used, companion skills, or follow-up actions when they materially improve execution.
Evidence Produced
| Category | Artifact | Format | Example |
|---|---|---|---|
| UX quality | AI surface design review | Markdown doc covering trustworthiness, error tolerance, and refusal-state handling for AI-powered features | docs/ai/ux-design-review.md |
References
- Use the links and companion skills already referenced in this file when deeper context is needed.
Based on Nudelman (2024) UX for AI: A Framework for Product Design.
When to Use
Load this skill when:
- Building any AI-powered feature, chatbot, or copilot
- Designing AI search, recommendations, anomaly detection, or generative outputs
- Reviewing AI features for trust, transparency, and premium feel
- Avoiding the "AI slop" feeling in AI-assisted products
1. Core Principles (RETCH)
Five non-negotiable principles for trustworthy AI UX:
Restate — Before acting, the AI must echo back what it understood. "Showing results for: [interpreted query]." This closes the gap between what the user meant and what AI understood without requiring re-entry.
Calibrate — AI must know what it doesn't know. Communicate confidence levels. When confidence is low, fall back to deterministic baselines or say "I don't know." An AI that cannot self-assess accuracy is a liability.
Explain — Show the "why" behind AI decisions. Citations, recommended reasons, translated queries. "Recommended because you viewed X." Users who understand AI reasoning trust it more and use it better.
Transparent — State what data the AI uses and whether conversations are stored. Don't bury privacy stance in documentation — state it directly in the AI interface.
Human oversight — Never remove manual override controls. Always provide stop, pause, and override buttons for any AI action. The Boeing 737 Max crash (346 deaths) is the foundational case study for what happens when AI silently overrides humans with no visible intervention path.
2. Mental Models Users Hold About AI
Users want outcomes, not AI. The mechanism is irrelevant. If a simpler deterministic solution delivers the outcome, use it. Never force AI into a product just because it is technically possible.
Algorithm aversion is real and manageable. Users tolerate bad AI output less than equivalent bad human output. However, aversion decreases when users can slightly modify the AI output, or when the system demonstrates it is improving over time.
AI must not claim to replace experts. Positioning AI as a direct replacement for an established expert (farmer, doctor, analyst) provokes distrust and resistance. AI should assist and augment the expert — never compete with their core skill.
Users expect contextual intelligence. Modern AI should infer context from calendar data, location, time of day, and prior conversation — without making users re-specify what they clearly mean.
Deskilling warning. Over-reliance on AI erodes the user's own capability. Design for augmentation, not replacement, especially in high-stakes domains (medicine, security, finance).
3. Making AI Feel Premium (Not Sloppy)
Domain-specific training is the single biggest differentiator. Generic ChatGPT answers feel wrong for your product. Fine-tune on your specific dataset or use RAG (retrieval-augmented generation). An AI that gives domain-accurate answers feels intelligent; one that gives generic LLM responses feels broken.
Stateful AI feels intelligent; stateless AI feels dumb. Allow users to resume investigations, maintain context across sessions, and build on prior interactions. Statelessness makes AI feel disconnected and forgetful.
Size the copilot to match task importance:
- Side panel → page-level, context-aware tasks (preferred)
- Large overlay → generally avoid; obscures parent page
- Full-page dedicated experience → deep, analytical, complex tasks
Context-aware initial suggestions feel like mind-reading. Surface the right suggestions before the user types anything — based on time of day, data type, prior sessions, and usage patterns. Generic "What can I help you with?" prompts feel cheap.
Use Promptbooks to reduce prompt engineering burden. Users should never need to write complex prompts. Provide pre-built recipes for common workflows. Curated, plain-language queries that interact with domain data in predictable ways is the target.
Give AI a deliberate personality. Define the persona explicitly (supportive health coach, expert security analyst, etc.). Avoid pedantic or preachy tone. Generic LLM personality is a missed differentiator.
Test with real AI output, not mocked text. Products that demo well on polished placeholder copy but fail with actual AI responses are caught in research, not production. Use Python notebooks + paper prototype ("Wizard of Oz") for honest testing.
4. Error States & Graceful Degradation
The core product must be resilient to bad AI output. AI autocomplete and pre-fill should save time when correct but must never block the user when wrong. Always preserve the manual fallback path.
Wrap all AI output in business logic guardrails. Never pass raw model output directly to users or downstream systems. Hard-code limits (max bid, medication dose, allowed actions) regardless of what the AI recommends.
The Value Matrix principle. False positive vs false negative costs are not equal. Calculate the real cost of each error type before choosing how to handle AI failures. Optimise for real-world ROI, not data science accuracy metrics. A "Balanced AI" regularly outperforms a "Accurate (Conservative) AI" in real-world ROI.
Progressive accuracy improvement is critical for retention. Users will abandon an AI product within 3–5 interactions if accuracy does not visibly improve with use. Design AI systems to learn from user corrections.
Guardrails layer required for all LLM outputs. Define explicitly what the AI will and will not output. Code explicit failure responses for denied or malformed requests. Guardrails must be designed, not improvised.
5. Onboarding AI Features
"Set it and forget it" for low-stakes AI. Activate with zero setup when the downside is minimal and immediate value is clear. Zoom AI Companion activating automatically on meeting start is the benchmark.
Lead with immediate, tangible value. The benefit must be clear in the first interaction. Users should not need to invest effort to see value.
Promptbooks and initial suggestions cure cold-start anxiety. An empty prompt box paralyses users. Offer context-aware suggestions based on data type, time of day, and prior sessions.
Validate the use case with storyboards first. A use case that sounds reasonable as text often falls apart when storyboarded step-by-step. The conclusion panel of the storyboard must follow naturally from the rest. If it doesn't, the use case is wrong.
Train on 2,000+ domain-specific examples before launch. An untrained LLM answering domain-specific queries feels broken. Users who encounter generic responses in their first interaction rarely return.
6. Progressive Disclosure for AI Capabilities
Start with the "one thing." Every AI feature needs a single, clear primary job. What is the one thing users should always be able to do with this AI? Everything else is secondary.
Reveal AI capabilities through contextual next steps, not documentation. After each AI interaction, surface the 2–3 most valuable next questions the user is likely to have. This teaches users what the AI can do — inside their actual workflow.
Use Mad Lib fill-in-the-blank for complex AI configuration. Write the configuration as a complete English sentence with embedded drop-downs: "Trigger a [Critical] alert whenever the value exceeds [90%] for [1 minute]" is far more comprehensible than a table of parameters.
Information architecture still matters in AI-first apps. Chat alone is not a complete interface. Users need predetermined starting points, context-aware alerts, and structured navigation. A blank chat box communicates nothing about what the product does.
7. Human-AI Collaboration Patterns
In agentic AI, the human's job is to accept or reject evidence, not direct every step. Design accept/reject flows for observations and hypotheses. The human is a reviewer and quality gate, not a director.
Agentic AI requires start, stop, pause, and approval controls. Agents are asynchronous and semi-autonomous. Without stop and pause controls, agents create runaway costs ("Sorcerer's Apprentice" problem). Approval gates for expensive or irreversible actions are mandatory.
Agents require RBAC, data isolation, and versioning. Treat AI agents like employees — with access controls, data isolation, and version management. This governance UI is a new and urgent design challenge.
Augmented intelligence, not replacement. Machines handle pattern recognition, data processing, and hypothesis generation. Humans handle empathy, ethics, judgment, and creative synthesis.
8. Anti-Patterns
- Replacing an expert directly — Always fails. AI should assist, not compete.
- Optimising on accuracy metrics alone — Real-world ROI is the only metric that matters.
- Large overlay copilot panels — Obscure parent content, force repeated open/close cycles. Use side panels instead.
- Chat-only information architecture — Chat is a command line, not a complete product.
- Copying ChatGPT patterns blindly — Invent patterns suited to your specific use case.
- Using synthetic users for AI research — AI research hallucinations miss contextual nuance. Real user research is non-optional.
- Assuming AI will behave correctly without UX oversight — The primary cause of automation-related disasters. Always design for failure.
- Fully automated report generation without human selection — For legally important documents, require humans to pin relevant data before AI writes the report.
9. Key Mantras
- "AI is too important to be left to data scientists." UX must be present from use case discovery through ethics review.
- "Never let the user leave empty-handed." Every AI interaction must end with the user holding something useful.
- "Accuracy optimised alone almost always underperforms AI that considers real-world costs."
- "If you cannot tell a compelling story, you don't have a chance." Storyboard every AI use case before building.
- "Now is not the time to copy — it is the time to invent." Find what is unique about your use case.
- "Recognition over recall." Even Boeing 737 Max pilots died partly because they had to remember a checklist under extreme stress. All AI error-recovery procedures must be on screen.
- "The best AI is augmented intelligence." Let machines handle what machines do best. Let humans handle the rest.
Sources
- Nudelman, G. (2024). UX for AI: A Framework for Product Design. Wiley.
More from peterbamuhigire/skills-web-dev
google-play-store-review
Google Play Store compliance and review readiness for Android apps. Use
77multi-tenant-saas-architecture
Use when designing or reviewing a multi-tenant SaaS platform — tenant
68jetpack-compose-ui
Jetpack Compose UI standards for beautiful, sleek, minimalistic Android
49gis-mapping
Use for web apps that need Leaflet-first GIS mapping, location selection,
49saas-accounting-system
Implement a complete double-entry accounting system inside any SaaS app.
47manual-guide
Generate end-user manuals and reference guides for ERP modules. Use when
40