replay-ux-research
Replay UX Research
Analyze session replays from real external users of sentry.io to surface UX patterns, pain points, and representative journeys for a given product area. This uses Sentry's own dogfooding org.
Inputs
$ARGUMENTS is the product area to research (e.g., "issues", "traces", "dashboards", "replays", "monitors", "releases", "alerts").
If $ARGUMENTS is empty, ask the user which product area to research.
Prerequisites
This skill requires the Sentry MCP server to be connected. The following tools are used:
search_events— Search for session replaysget_replay_details— Get detailed replay informationsearch_issues— Look up error issuesget_sentry_resource— Fetch issue details from URLs
If these tools are not available, ask the user to connect the Sentry MCP server before proceeding.
Step 1: Map product area to URL patterns
Read references/product-areas.md and find the URL patterns for the requested area.
If the product area is not listed, infer a URL pattern from the area name. Most Sentry product areas follow the pattern /<area-name>/ in the URL path. The reference file may not cover newer product areas — confirm your assumption with the user if unclear.
Step 2: Search for replays
Search for replays from external (non-Sentry-employee) users. 25 replays is a good starting point — go deeper if the product area is complex, if early patterns are ambiguous, or if the user wants a more comprehensive picture.
Start with last 24 hours — extend to 48h or 7d if needed to reach your target count. Run multiple search_events calls if needed. Use limit: 50 per call.
If you can't find enough replays (fewer than 10 even at 7 days), tell the user what you found and ask them to help iterate — they may suggest broader URL patterns, a different time range, or a related product area to include.
Query construction:
Use natural language queries like:
replays from the last 24 hours where url contains "/<area-path>" excluding user emails ending in @sentry.io and @getsentry.com
Key filters:
- Time range: last 24 hours (extend if needed)
- URL pattern: match the product area paths from Step 1
- Exclude employees:
-user.email:*@sentry.io -user.email:*@getsentry.com - Environment: prod
Do NOT pass a projectSlug filter — replays span the whole org.
Step 3: Get replay details
Call get_replay_details for each replay found in Step 2. Run these calls in parallel batches for speed.
For each replay, capture:
- User: email domain only (the API returns full emails — never include these in output)
- Journey: ordered list of pages visited (from URLs and activity breadcrumbs)
- Duration: total session length
- Replay type:
session(randomly sampled from normal browsing) vsbuffer(triggered by an event — error, manual flush, or specific user action like submitting feedback or going through checkout). Note this distinction in your analysis since buffer replays are biased toward error/action moments, not typical browsing. - Entry context: first URL tells you how they arrived — look for referrer signals like
referrer=slack,notification_uuid,alert_rule_idin query params (Slack notification), email link patterns, or bare URLs (bookmark/direct nav) - Engagement signals: error count, rage clicks, dead clicks, warning count
- Browser/OS/Device: for device distribution context
- Activity breadcrumbs: page views, navigation patterns, key interactions
Step 4: Investigate significant errors
After collecting replay details, identify errors that appear in multiple replays or seem likely to affect the user experience. For each significant error:
-
Triage by frequency: If the same issue ID (e.g.,
JAVASCRIPT-33RM) appears in 3+ replays, it's worth investigating. -
Check the issue in Sentry: Use
search_issuesto find the issue, orget_sentry_resourcewith the issue URL from the replay details. Understand:- What is the error? (message, stack trace context)
- How many total users/events does it affect? (beyond just this replay sample)
- Is it assigned or being worked on?
-
Infer user-facing impact from behavioral signals: We cannot see the rendered page content through replay metadata — only by watching the replay in-browser. Instead, infer impact from what users did after the error:
- Retried the same action → they likely saw a failure and tried again
- Navigated away immediately → they were blocked or gave up
- Continued their flow normally → the error may be silent/cosmetic
- Rage-clicked or dead-clicked after → the UI may have become unresponsive
- Spent a long time on the page after → they may be reading an error message or confused
- No behavioral change at all → error was likely invisible to the user
-
Classify each error based on this evidence:
- Likely blocking: Error + user retried/left/couldn't continue. High confidence of user impact.
- Likely degrading: Error + user continued but with unusual behavior. Moderate confidence.
- Likely silent: Error fired but user behavior was unaffected. Low confidence of user impact.
- Unclear: Not enough behavioral signal to judge. Flag for manual replay review.
Always note the confidence level and recommend watching specific replays to confirm. Link directly to the replay URL for each classified error.
Include this classification in the Friction & Pain Points section. Don't report likely-silent errors as pain points — list them in a separate "Background Errors (likely silent)" subsection for completeness.
Step 5: Analyze patterns
Look at the replays through these UX research lenses:
Behavioral patterns
- Common journeys: What navigation paths do users take? What's the typical flow?
- Entry points: How do users arrive? Categorize: alert notification (Slack/email), direct bookmark, organic navigation from another page. The first URL's query params reveal this.
- Task completion: Did the user appear to accomplish a goal, or did they wander/abandon? Signs of completion: navigating to a detail view then leaving. Signs of abandonment: short session, back-and-forth navigation, leaving from the same page they entered.
- Time on task: How long do users spend on key pages before acting?
Friction & discovery
- Friction signals: Rage clicks, dead clicks, errors — but also hesitation (visiting the same page repeatedly), thrashing (rapid back-and-forth between pages), and retry loops (repeating the same action sequence).
- Feature discovery: Are users finding sub-features (filters, search, sort, bulk actions) or only using the primary view? Look at URL query params and breadcrumbs for evidence of feature use.
- User intent signals: Mine URL query params for search terms, filter values, sort orders, and date ranges users set. These are the closest thing to verbatim user "quotes" — they reveal what users are looking for in their own words. (e.g.,
query=is%3Aunresolved+assigned%3Ametells you the user is triaging their own assigned issues.) - Workarounds: Any unexpected navigation patterns that suggest a missing feature or confusing flow? (e.g., going to settings mid-task, opening multiple pages in sequence that could be one view)
- Error recovery: When users encounter errors, do they recover and continue or abandon?
Context
- Replay trigger mix: What proportion are
session(random sample) vsbuffer(event-triggered)? Buffer replays show moments where something notable happened (error, feedback submission, checkout, etc.) — they're valuable for friction analysis but aren't representative of typical browsing. Call out this bias when drawing conclusions. - Return visitors: Do any email domains appear in multiple replays? Repeat visitors suggest habitual usage — their journeys may reveal power-user patterns or persistent pain points they've learned to work around.
- User diversity: Are replays from many different orgs/companies or concentrated? Are there differences in behavior by org?
- Device/browser distribution: What are users primarily using?
- Drop-off points: Where do users leave or navigate away?
Step 6: Write the report
Use the template in references/output-template.md. Be specific — cite individual replays as evidence for each pattern. Link to replay URLs so the reader can watch the replay themselves.
Privacy: Never include full user email addresses in the report. Use anonymized identifiers like "user from [company domain]" or "User A, B, C."
More from getsentry/skills
security-review
Security code review for vulnerabilities. Use when asked to "security review", "find vulnerabilities", "check for security issues", "audit security", "OWASP review", or review code for injection, XSS, authentication, authorization, cryptography issues. Provides systematic review with confidence-based reporting.
4.9Kcode-simplifier
Simplifies and refines code for clarity, consistency, and maintainability while preserving all functionality. Use when asked to "simplify code", "clean up code", "refactor for clarity", "improve readability", or review recently modified code for elegance. Focuses on project-specific best practices.
4.1Kfind-bugs
Find bugs, security vulnerabilities, and code quality issues in local branch changes. Use when asked to review changes, find bugs, security review, or audit code on the current branch.
2.2Kagents-md
Creates and maintains concise AGENTS.md and CLAUDE.md project instruction files. Use when asked to create AGENTS.md, update AGENTS.md, maintain agent docs, set up CLAUDE.md, document repository agent conventions, or keep coding-agent instructions minimal and reference-backed.
2.1Kcode-review
Perform code reviews following Sentry engineering practices. Use when reviewing pull requests, examining code changes, or providing feedback on code quality. Covers security, performance, testing, and design review.
2.0Kcommit
ALWAYS use this skill when committing code changes — never commit directly without it. Creates commits following Sentry conventions with proper conventional commit format and issue references. Trigger on any commit, git commit, save changes, or commit message task.
1.9K