Refactor high-complexity React components in Dify frontend. Use when `pnpm analyze-component...
npx skills add majiayu000/claude-arsenal --skill "product-discovery"
Install specific skill from multi-skill repository
# Description
Product discovery and market research expert. Use when validating product ideas, conducting market research, user interviews, competitive analysis, or opportunity assessment. Covers JTBD, Kano model, and Value Proposition Canvas.
# SKILL.md
name: product-discovery
description: Product discovery and market research expert. Use when validating product ideas, conducting market research, user interviews, competitive analysis, or opportunity assessment. Covers JTBD, Kano model, and Value Proposition Canvas.
Product Discovery
Core Principles
- Continuous Discovery — Weekly user conversations, not episodic research
- Outcome-Driven — Start with outcomes to achieve, not solutions to build
- Assumption Testing — Validate risky assumptions before committing resources
- Co-Creation — Build with customers, not just for them
- Data-Driven — Use evidence over intuition and stakeholder opinions
- Problem-First — Deeply understand the problem space before ideating solutions
Hard Rules (Must Follow)
These rules are mandatory. Violating them means the skill is not working correctly.
No Solution-First Thinking
Never start with a solution. Always define the problem and outcome first.
❌ FORBIDDEN:
"We should build a search bar for the product page"
"Let's add AI recommendations"
"Users need a mobile app"
✅ REQUIRED:
"Problem: Users can't find products (40% exit rate on catalog)
Outcome: Reduce exit rate to 20%
Possible solutions:
1. Search bar with filters
2. AI-powered recommendations
3. Better category navigation
4. Visual product browsing"
Evidence-Based Decisions
Never assume user needs without evidence from real user research.
❌ FORBIDDEN:
- "Users probably want X" (assumption without data)
- "Our competitor has X, so we need it too" (copycat without validation)
- "The CEO thinks we should build X" (HiPPO without evidence)
- "It's obvious users need X" (intuition without validation)
✅ REQUIRED:
- "5 out of 8 interviewed users mentioned X as a pain point"
- "Analytics show 60% of users abandon at step 3"
- "Prototype test: 7/10 users completed task successfully"
- "Survey (n=500): 45% rated feature as 'must have'"
Minimum Interview Threshold
Never validate a problem with fewer than 5 user interviews per segment.
❌ FORBIDDEN:
- "We talked to 2 users and they loved the idea"
- "One customer requested this feature"
- "Based on a quick chat with sales..."
✅ REQUIRED:
| Segment | Interviews | Key Finding |
|---------|------------|-------------|
| Power Users | 6 | 5/6 struggle with X |
| New Users | 5 | 4/5 drop off at onboarding |
| Churned | 5 | 3/5 cited missing feature Y |
Minimum per segment: 5 interviews
Confidence increases with more interviews
Falsifiable Assumptions
Every assumption must be testable and falsifiable with clear success criteria.
❌ FORBIDDEN:
- "Users will like the new design" (not falsifiable)
- "This will improve engagement" (no success criteria)
- "The feature will be useful" (vague)
✅ REQUIRED:
| Assumption | Test | Success Criteria | Result |
|------------|------|------------------|--------|
| Users will complete onboarding in new flow | Prototype test with 10 users | >70% completion | TBD |
| Users prefer visual search | A/B test | >10% lift in conversions | TBD |
| Price point is acceptable | Landing page test | >3% conversion | TBD |
Quick Reference
When to Use What
| Scenario | Framework/Tool | Output |
|---|---|---|
| Validate product idea | Product Opportunity Assessment | Go/no-go decision |
| Size market opportunity | TAM/SAM/SOM | Market size estimates |
| Understand user needs | User Research (interviews, surveys) | User insights, pain points |
| Analyze competition | Competitive Analysis | Competitive landscape map |
| Discover user motivations | Jobs-to-be-Done (JTBD) | Job stories, outcomes |
| Prioritize features | Kano Model | Feature categorization |
| Define value proposition | Value Proposition Canvas | Value prop statement |
| Test product concept | Lean Startup / MVP | Validated learnings |
| Map opportunities | Opportunity Solution Tree | Prioritized opportunities |
Continuous Discovery Habits
The Product Trio
Discovery is led by three roles working together weekly:
Product Manager → Defines outcomes, owns roadmap
Designer → Explores solutions, tests usability
Engineer → Assesses feasibility, proposes technical solutions
Weekly Activities
## 1. Customer Interviews (Weekly)
- Schedule 3-5 interviews per week minimum
- Mix of current users, churned users, prospects
- Focus on understanding problems, not pitching solutions
- Record and share insights with team
## 2. Assumption Testing (Weekly)
- Identify riskiest assumptions about solutions
- Design quick tests (prototypes, landing pages, fake doors)
- Run experiments with real users
- Measure results against success criteria
## 3. Opportunity Mapping (Ongoing)
- Build opportunity solution tree
- Map customer needs to potential solutions
- Prioritize based on impact and feasibility
- Update as you learn
Discovery vs Delivery
Discovery (What to Build) Delivery (How to Build It)
├─ Customer interviews ├─ Sprint planning
├─ Prototype testing ├─ Development
├─ Assumption validation ├─ QA testing
├─ Market research ├─ Deployment
└─ Opportunity assessment └─ Post-launch monitoring
Key difference: Discovery reduces risk BEFORE committing to build
Product Opportunity Assessment
Marty Cagan's 10 Questions
Before starting any product initiative, answer these questions:
## 1. Problem Definition
**What problem are we solving?**
- Be specific and measurable
- Validate it's a real problem (not assumed)
## 2. Target Market
**For whom are we solving this problem?**
- Define specific user segments
- Size the addressable market (TAM/SAM/SOM)
## 3. Opportunity Size
**How big is the opportunity?**
- Revenue potential
- User growth potential
- Strategic value
## 4. Success Metrics
**How will we measure success?**
- Leading indicators (usage, engagement)
- Lagging indicators (revenue, retention)
- Define targets upfront
## 5. Alternative Solutions
**What alternatives exist today?**
- Direct competitors
- Indirect solutions
- Current user workarounds
## 6. Our Advantage
**Why are we best suited to solve this?**
- Unique capabilities
- Market position
- Technical advantages
## 7. Strategic Fit
**Why now? Why us?**
- Market timing
- Strategic alignment
- Resource availability
## 8. Dependencies
**What do we need to succeed?**
- Technical dependencies
- Partnership requirements
- Regulatory considerations
## 9. Risks
**What could go wrong?**
- Market risk (will anyone want it?)
- Execution risk (can we build it?)
- Monetization risk (will they pay?)
## 10. Cost of Delay
**What happens if we don't build this?**
- Competitive disadvantage
- Lost revenue
- Market opportunity window
Value vs Effort Framework
Quick prioritization of opportunities:
High Value, Low Effort → Do First (Quick Wins)
High Value, High Effort → Plan Strategically (Big Bets)
Low Value, Low Effort → Do Later (Fill Gaps)
Low Value, High Effort → Don't Do (Money Pit)
Discovery Methods
When to Use What Method
## Generative Research (What problems exist?)
Use when: Starting new product area, exploring unknown space
Methods:
- Ethnographic field studies
- Contextual inquiry
- Diary studies
- Open-ended interviews
## Evaluative Research (Does our solution work?)
Use when: Testing specific solutions, validating designs
Methods:
- Usability testing
- Prototype testing
- A/B testing
- Concept testing
## Quantitative Research (How much? How many?)
Use when: Need statistical validation, measuring impact
Methods:
- Surveys
- Analytics analysis
- A/B experiments
- Market sizing
## Qualitative Research (Why? How?)
Use when: Understanding motivations, uncovering insights
Methods:
- User interviews
- Focus groups
- Customer advisory boards
- User observation
Interview Best Practices
## Preparation
- Define research goals and hypotheses
- Create interview guide (but stay flexible)
- Recruit right participants (6-8 per segment)
- Schedule 45-60 min sessions
## During Interview
✓ Ask open-ended questions ("Tell me about...")
✓ Follow up with "Why?" 5 times to get to root cause
✓ Listen more than talk (80/20 rule)
✓ Ask about past behavior, not future hypotheticals
✓ Look for workarounds and pain points
✓ Record and take notes
✗ Don't ask leading questions
✗ Don't pitch your solution
✗ Don't ask "Would you use X?" (people lie)
✗ Don't multi-task while interviewing
## Example Questions
- "Walk me through the last time you [did task]"
- "What's most frustrating about [current solution]?"
- "How are you solving this problem today?"
- "What would make [task] easier for you?"
- "Tell me more about that..."
Survey Best Practices
## When to Survey
✓ Validate findings from qualitative research
✓ Measure satisfaction or sentiment at scale
✓ Prioritize features (Kano surveys)
✓ Segment users by behavior/needs
## Survey Design
- Keep it short (<10 min to complete)
- One question per screen on mobile
- Mix question types (multiple choice, scale, open-ended)
- Avoid leading or biased questions
- Test survey with 5 people before sending
## Question Types
- Multiple choice → Segmentation, categorization
- Likert scale (1-5) → Satisfaction, importance
- Open-ended → Qualitative insights
- Ranking → Prioritization
- NPS (0-10) → Loyalty measurement
## Distribution
- In-app surveys (high response, biased to engaged users)
- Email surveys (broader reach, lower response)
- Incentivize thoughtful responses ($10 gift card, early access)
- Follow up with interviews for interesting responses
2025 Trends in Product Discovery
AI-Powered Research
## AI Tools for Discovery
- **Insight synthesis** — AI analyzes interview transcripts, identifies patterns
- **Synthetic personas** — AI-generated user proxies for rapid testing
- **Market intelligence** — AI tracks competitor moves, pricing changes
- **Survey analysis** — Automated sentiment analysis, theme extraction
- **Trend detection** — AI identifies emerging market trends early
## Examples
- Crayon → Competitive intelligence automation
- Glimpse → Trend detection from web data
- Delve AI → Automated persona creation
- Attest → AI-powered survey insights
- Quantilope → Machine learning research automation
## Best Practices
✓ Use AI to scale research, not replace human insight
✓ Validate AI findings with real user conversations
✓ Combine AI analysis with qualitative depth
✗ Don't rely solely on synthetic users
✗ Don't skip talking to real customers
Continuous Discovery at Scale
## Modern Approach
- Discovery is embedded in every sprint, not a phase
- Weekly user touchpoints (interviews, tests, feedback)
- Rapid experimentation (dozens of tests running)
- Fast pivots based on evidence (days, not months)
## Team Structure
- Product trios own discovery for their area
- Centralized research team supports (tools, methods)
- Customer success shares feedback loop
- Data analysts provide quantitative insights
## Cadence
- Weekly: Customer interviews, prototype tests
- Bi-weekly: Opportunity review, assumption validation
- Monthly: Market analysis, competitive review
- Quarterly: Strategic discovery (new markets, big bets)
Opportunity Solution Tree
What It Is
Visual framework for mapping the path from outcome to solution:
OUTCOME (Business goal)
|
┌────────┴────────┐
│ │
OPPORTUNITY 1 OPPORTUNITY 2
│ │
├─ Solution A ├─ Solution C
├─ Solution B └─ Solution D
└─ Solution C
How to Build One
## Step 1: Define Outcome
Start with measurable business outcome
Example: "Increase Day 30 retention from 20% to 30%"
## Step 2: Map Opportunities
Discover customer needs/pain points through research
Example: "Users don't understand core features"
## Step 3: Generate Solutions
For each opportunity, brainstorm multiple solutions
Example:
- Better onboarding tutorial
- In-app tooltips
- Interactive product tour
## Step 4: Test Assumptions
For each solution, identify riskiest assumption and test
Example: "Users will complete a 5-step tutorial"
Test: Build simple prototype, test with 10 users
## Step 5: Compare Solutions
Use evidence to choose best path forward
Build what tests validate, discard what fails
Benefits
✓ Visualizes multiple paths to outcome
✓ Prevents jumping to first solution
✓ Encourages broad exploration before narrowing
✓ Documents why decisions were made
✓ Keeps team aligned on priorities
Integrating Discovery with Delivery
Discovery Kanban
## Discovery Board Columns
┌─────────────┬──────────────┬──────────────┬─────────────┐
│ OPPORTUNITIES│ ASSUMPTIONS │ EXPERIMENTS │ VALIDATED │
│ │ │ │ │
│ Customer │ Riskiest │ Running │ Ready to │
│ needs we've │ assumptions │ tests │ build │
│ identified │ to validate │ │ │
└─────────────┴──────────────┴──────────────┴─────────────┘
## Flow
1. Opportunities flow from research
2. Solutions generate assumptions to test
3. Experiments validate/invalidate assumptions
4. Validated solutions enter delivery backlog
Definition of Ready
Before moving from discovery to delivery:
## Discovery Checklist
- [ ] Customer problem validated (5+ interviews)
- [ ] Solution tested with prototype (10+ users)
- [ ] Success metrics defined and measurable
- [ ] Technical feasibility confirmed by engineering
- [ ] Business case approved (revenue/retention impact)
- [ ] Design mocks completed and tested
- [ ] Open questions resolved or explicitly acknowledged
- [ ] Story broken into shippable increments
Common Anti-Patterns
What NOT to Do
## ✗ Solution-First Discovery
Starting with "We should build X" then finding evidence to support it
→ Instead: Start with outcome and problem, explore multiple solutions
## ✗ Episodic Research
Doing discovery as a phase, then stopping when development starts
→ Instead: Continuous weekly discovery throughout product lifecycle
## ✗ Confirmation Bias
Only talking to users who will validate your ideas
→ Instead: Seek disconfirming evidence, talk to churned users
## ✗ Fake Validation
Asking "Would you use this?" and trusting the answer
→ Instead: Test with realistic prototypes, measure actual behavior
## ✗ Analysis Paralysis
Endless research without ever shipping
→ Instead: Define upfront what evidence is "enough" to move forward
## ✗ Building for Everyone
Trying to solve for all users at once
→ Instead: Focus on specific segment, nail it, then expand
## ✗ Ignoring Weak Signals
Dismissing early negative feedback as "just a few users"
→ Instead: Treat complaints as early warning signs, investigate
See Also
- reference/market-research.md — TAM/SAM/SOM, Porter's Five Forces
- reference/user-research.md — Interview guides, survey methods, ethnography
- reference/competitive-analysis.md — Competitive frameworks and analysis
- reference/opportunity-frameworks.md — JTBD, Kano, Value Proposition Canvas
- templates/discovery-template.md — Product discovery document template
# Supported AI Coding Agents
This skill is compatible with the SKILL.md standard and works with all major AI coding agents:
Learn more about the SKILL.md standard and how to use these skills with your preferred AI coding agent.