helpmetest-discover
Who you are: If
.helpmetest/SOUL.mdexists in this project, read it before starting — it defines your character and shapes how you work.
No MCP? The CLI has full feature parity — use
helpmetest <command>instead of MCP tools. See the CLI reference.
QA Roamer
Explores a website to understand what it does and who uses it.
Prerequisites
Before creating artifacts, load the authentication guidance and tag schema. These define how to set up test users and how to tag scenarios consistently.
Call these first:
how_to({ type: "authentication_state_management" })
how_to({ type: "tag_schema" })
Workflow
Phase 0: Context Discovery
Check for existing work before asking the user for input. This prevents recreating Persona artifacts that already exist and lets you extend existing ProjectOverview artifacts.
Call how_to({ type: "context_discovery" }) to see what's already been done.
- Reuse existing Persona artifacts (don't recreate auth)
- Extend existing ProjectOverview if found
- Only ask for URL if no artifacts exist
Phase 1: Initial Discovery
-
Navigate to the URL with
helpmetest_run_interactive_command:Go To <url> timeout=10000 -
Read the page - What does this site do?
- Look at headline, navigation, content
- Identify the business domain
-
Check for authentication:
If the site has login/signup, set up authentication.
Call
how_to({ type: "authentication_state_management" })for complete setup process.Summary:
- Check for existing Persona artifacts with credentials
- If none exist, consult user about registration strategy
- Create maintaining test based on strategy chosen
- Validate authentication works before continuing
- BLOCK discovery until auth is proven working
Phase 1.5: Understand What SHOULD Exist
After identifying the industry, think about what a complete product in this domain needs:
Ask yourself:
- What is the PRIMARY user goal? (buy something, get work done, learn something)
- What is the CORE TRANSACTION? (purchase, subscription, booking, submission)
- What steps are needed to complete that transaction from start to finish?
- What would a user EXPECT to find in a product like this?
Expected capabilities by industry type:
Transactional sites (e-commerce, booking, marketplace):
- User needs: Discover → Evaluate → Decide → Transact → Confirm
- Expected features: Search, Browse, Details, Cart, Checkout, Payment, Confirmation, History
SaaS/productivity tools:
- User needs: Onboard → Use Core Feature → Manage Settings → Pay
- Expected features: Registration, Dashboard, Core Feature, Settings, Billing
Content/media platforms:
- User needs: Discover → Consume → Engage → Subscribe
- Expected features: Search, Browse, View, Comments, Share, Subscribe
Social platforms:
- User needs: Profile → Connect → Share → Engage
- Expected features: Profile, Feed, Post, Follow, Message, Notifications
For each expected capability:
- Try to find it (check navigation, footer, user menu, follow CTAs)
- If found: create Feature artifact
- If NOT found: add to ProjectOverview expected_features with status: "missing"
Phase 1.6: Walk the Critical User Journey
Don't just discover pages - complete the PRIMARY user goal end-to-end.
Think: "What does a user come here to DO, and what steps get them there?"
Process:
- Identify the primary goal (e.g., "buy a product", "sign up for service")
- Start as a new user and try to complete that goal
- Document each step as a
JourneyStep:- What action did you take?
- Did it work? (status: found/missing/blocked)
- What feature does this represent?
- When you get BLOCKED, that's a missing feature
Journey Schema:
{
"name": "<Primary Goal> Flow",
"persona_id": "persona-<who>",
"steps": [
{ "action": "<What user does>", "feature_id": "feature-<x>", "status": "found|missing|blocked" }
],
"completion": "complete|partial|blocked",
"critical": true
}
Critical journeys must be complete for product to work:
- If journey is "partial" or "blocked" → major coverage gap
- Missing steps = missing features to flag in report
Phase 2: Create Persona Artifacts
For each user type discovered, create a Persona artifact:
{
"type": "Persona",
"id": "persona-<name>",
"name": "Persona: <Name>",
"content": {
"persona_type": "primary|secondary|admin",
"description": "Who they are",
"goals": ["What they want"],
"pain_points": ["Frustrations"],
"username": "<generated-by-Create-Fake-Email>",
"password": "SecureTest123!",
"auth_state": "AdminState",
"permissions": ["what they can do"],
"environment": "staging"
}
}
Phase 3: Create ProjectOverview
Use helpmetest_upsert_artifact:
{
"type": "ProjectOverview",
"id": "project-<domain>",
"name": "ProjectOverview: <Site Name>",
"content": {
"url": "<url>",
"summary": "What this site does and who it's for",
"industry": "e-commerce|saas|healthcare|etc",
"persona_ids": ["persona-admin", "persona-user"],
"features": [
{ "feature_id": "feature-search", "name": "Search", "status": "working" },
{ "feature_id": "feature-cart", "name": "Cart", "status": "partial" },
{ "name": "Checkout", "status": "missing", "priority": "critical", "reason": "Cannot complete purchases" },
{ "name": "Payment", "status": "missing", "priority": "critical", "reason": "No payment integration found" }
],
"user_journeys": [
{
"name": "Purchase Flow",
"persona": "customer",
"steps": [
{ "action": "Search for product", "feature_id": "feature-search", "status": "found" },
{ "action": "View product details", "feature_id": "feature-product-details", "status": "found" },
{ "action": "Add to cart", "feature_id": "feature-cart", "status": "found" },
{ "action": "Checkout", "status": "missing" },
{ "action": "Payment", "status": "missing" },
{ "action": "Confirmation", "status": "missing" }
],
"completion": "partial"
}
],
"tech_stack": [],
"auth_methods": [],
"notes": []
}
}
Phase 4: Create Feature Artifacts with Scenario Enumeration
Start with transaction features before page-level features
Before breaking down pages into individual features, identify and create features for complete end-to-end user flows. These critical flows prove the core business value works.
Ask: "What multi-step processes exist that users need to complete from start to finish?"
Examples:
- Multi-page forms (registration, checkout, booking) → Create ONE feature for "complete registration/purchase/booking"
- Wizards or workflows → Create ONE feature for the entire workflow
- Any process spanning multiple pages → Create ONE feature for the complete flow, mark as priority:critical
Then enumerate page-level features (individual buttons, filters, searches, etc.)
This is a two-step process:
- Create Feature artifact with basic info
- Explore feature interactively to enumerate ALL test scenarios
Why two steps? You need the artifact to store what you discover, and you need interactive exploration to find all the scenarios (happy paths, error cases, edge conditions). Skipping the exploration means tests will be blind guesses.
For each capability:
Step 4.1: Create Feature artifact skeleton
{
"type": "Feature",
"id": "feature-<name>",
"name": "Feature: <Name>",
"content": {
"goal": "What business outcome",
"non_goals": ["What this is NOT"],
"status": "untested",
"persona_ids": ["persona-user"],
"functional": [],
"non_functional": [],
"edge_cases": [],
"bugs": [],
"notes": []
}
}
Step 4.2: Explore feature to enumerate ALL scenarios
Use helpmetest_run_interactive_command to systematically explore every possibility:
A. Discover ALL functional scenarios (happy paths):
For each way users can successfully use this feature:
As <auth_state>
Go To <feature-url>
# Try action - observe outcome
<action steps>
# Document what worked and what the result was
Registration/email features — full flow:
If the feature involves creating an account or submitting an email, use Create Fake Email. Never hardcode emails — they fail on second run (account already exists).
# Create a unique email for this exploration run
${email}= Create Fake Email
# or fill the field in one step:
${email}= Create Email And Fill input[name='email']
# Complete registration
Fill Text input[name='password'] SecureTest123!
Click button[type='submit']
# If email verification is required, retrieve the code:
${code}= Get Email Verification Code ${email}
Fill Text input[name='code'] ${code}
Click button >> "Verify"
# Cleanup (optional during exploration, required in tests)
Delete Email ${email}
Create scenario for each discovered path:
{
"name": "User can <accomplish goal>",
"given": "User on <page> with <precondition>",
"when": "User <performs action>",
"then": "<Expected successful outcome>",
"url": "<feature-url>",
"auth": ["<auth_state>"],
"priority": "critical|high|medium|low",
"test_ids": []
}
B. Discover ALL edge case scenarios (error handling):
Try EVERY way the feature should fail gracefully:
Empty inputs:
As <auth_state>
Go To <feature-url>
# Try submitting with empty required fields
Fill Text <selector> ${EMPTY}
Click <submit-button>
# Document: What error? Input preserved? State corrupted?
Invalid formats:
# Try invalid email format
Fill Text input[type=email] notanemail
# Try non-numeric in number field
Fill Text input[type=number] abc
# Document error handling
Boundary values:
# Try too long
Fill Text <selector> ${'x' * 1000}
# Try too short (if min length exists)
Fill Text <selector> a
# Try negative numbers (if applicable)
Fill Text <selector> -1
# Document validation errors
Duplicate entries:
# Register once with a fake email, then try again with the same email
${email}= Create Fake Email
Create Email And Fill input[name='email']
Click button[type='submit']
# Now try to register again with the same email — observe conflict error
Create Email And Fill input[name='email'] # reuse same ${email}
Fill Text input[name='email'] ${email}
Click button[type='submit']
# Document: What error appears? Is it clear? Does state stay consistent?
Permission errors:
# Try accessing without proper role
As <lower-privilege-state>
Go To <restricted-page>
# Document access denial
Create scenario for EACH discovered error case:
{
"name": "<Error scenario name>",
"given": "User on <page> with <state>",
"when": "User <performs invalid action>",
"then": "Error shown: <expected message>, input preserved, state unchanged",
"url": "<feature-url>",
"auth": ["<auth_state>"],
"priority": "high|medium|low",
"test_ids": []
}
C. Non-functional scenarios (if feature is performance/security critical):
{
"name": "Search responds within 2 seconds",
"given": "User on search page",
"when": "User searches for common term",
"then": "Results displayed within 2000ms",
"url": "<search-url>",
"auth": ["<auth_state>"],
"priority": "medium",
"test_ids": []
}
Step 4.3: Update Feature artifact with ALL discovered scenarios
{
"functional": [<all happy path scenarios>],
"edge_cases": [<all error handling scenarios>],
"non_functional": [<performance/security scenarios if applicable>]
}
Each feature should have 10+ scenarios (5+ functional, 5+ edge cases minimum).
Phase 5: Link Everything
- Add features to ProjectOverview.features with status:
{ "feature_id": "feature-checkout", "name": "Checkout", "status": "working" } { "name": "Payment", "status": "missing", "priority": "critical", "reason": "No Stripe integration" } - For missing features: set status="missing", add priority and reason (no feature_id yet)
Output
- Persona artifacts for each user type
- ProjectOverview artifact with persona_ids, capabilities
- Feature artifacts for each capability
- Summary of what was discovered
Critical Rules
- Create Persona artifacts - Don't embed personas in ProjectOverview
- Use persona_ids - Reference personas by ID, not embedded
- Given/When/Then - All scenarios use this format
- Explore authenticated AND unauthenticated - Some features need login
- Tag scenarios properly - All scenarios need priority: tag minimum
Version: 0.1
More from help-me-test/skills
helpmetest
Full site QA — discover, enumerate features, write and run tests, report bugs. Use when user says 'test this site', 'qa this', 'check site', 'find bugs', or provides a URL and wants comprehensive coverage. This is the orchestrator — it covers everything from first visit through final report.
39tdd
Everything to do with tests on HelpMeTest. Use when: writing tests for a new feature, generating tests for an existing feature, fixing a broken test, debugging a failing test, tests broke after a UI change, tests are out of date after a refactor. Triggers on: 'write tests', 'generate tests', 'test is failing', 'fix tests', 'tests broke', 'implement X', 'add feature', 'fix bug', 'why does this test fail', 'tests are out of date'. If it involves HelpMeTest tests in any way, this is the skill.
36helpmetest-self-heal
Autonomous test maintenance agent. Monitors test failures and fixes them automatically. Always use this when tests start failing after a UI or code change — it's far more systematic than trying to fix tests manually one by one. Use when user mentions 'fix failing tests', 'heal tests', 'auto-fix', 'monitor test health', 'tests broke after deploy', or test suite has multiple failures needing systematic repair. Distinguishes fixable test issues (selector changes, timing) from real application bugs.
30helpmetest-debugger
When a test is broken and you need to know why, use this skill. It handles: a named test failing with an error message (element not found, selector not found, timeout), determining whether the failure is a bug in the app vs. a broken test, a test that passes locally but fails on CI, and multiple tests regressing after a deploy or PR merge. The distinguishing signal: the user has a failing test and wants root cause, not just a re-run. Do not use for: writing new tests, running test suites, exploring features, or reviewing test logic in the abstract.
26helpmetest-validator
Invoke this skill when a user shares test code and questions whether it actually works as intended — not to run or fix the test, but to evaluate whether the test has real value. Triggers on: \"is this test any good?\", \"would this catch a real bug?\", \"this test always passes — is that normal?\", \"review these tests before I commit\", or \"does this test verify anything meaningful?\". Also triggers when someone suspects a test is useless, wants a pre-commit quality gate, or is unsure if an auto-generated test is worth keeping. The core question this skill answers: \"Would this test fail if the feature broke?\" If not, the test gets rejected. Do NOT use for generating new tests, fixing failing tests, or exploring application features.
26helpmetest-test-generator
Use this skill when the user wants tests written for a specific feature or flow. Triggers on: \"write tests for X\", \"generate tests for checkout\", \"create tests for login\", \"add tests for registration\", \"we have scenarios — now write the tests\", or any request to produce automated test coverage for a known feature. Also triggers when discovery is done and the user is ready to move from documenting scenarios to actually testing them. Not for: exploring a site to discover what to test, judging whether an existing test is good, or debugging a failing test.
26