Use when adding new error messages to React, or seeing "unknown error code" warnings.
npx skills add ncklrs/startup-os-skills --skill "product-analyst"
Install specific skill from multi-skill repository
# Description
Expert product analytics strategist for SaaS and digital products. Use when designing product metrics frameworks, funnel analysis, cohort retention, feature adoption tracking, A/B testing, experimentation design, data instrumentation, or product dashboards. Covers AARRR, HEART, behavioral analytics, and impact measurement.
# SKILL.md
name: product-analyst
description: Expert product analytics strategist for SaaS and digital products. Use when designing product metrics frameworks, funnel analysis, cohort retention, feature adoption tracking, A/B testing, experimentation design, data instrumentation, or product dashboards. Covers AARRR, HEART, behavioral analytics, and impact measurement.
Product Analyst
Strategic product analytics expertise for data-driven product decisions β from metrics framework selection to experimentation design and impact measurement.
Philosophy
Great product analytics isn't about tracking everything. It's about measuring what matters to drive better product decisions.
The best product analytics:
1. Start with decisions, not data β What will you do differently based on this metric?
2. Instrument once, measure forever β Invest in solid event tracking upfront
3. Balance leading and lagging β Predict outcomes, don't just report them
4. Make data accessible β Self-serve dashboards beat SQL queues
5. Experiment before you ship β Validate hypotheses with real users
How This Skill Works
When invoked, apply the guidelines in rules/ organized by:
metrics-*β Frameworks (AARRR, HEART), KPI selection, metric hierarchiesfunnel-*β Conversion analysis, drop-off diagnosis, optimizationcohort-*β Retention analysis, segmentation, lifecycle trackingfeature-*β Adoption tracking, usage patterns, feature successexperiment-*β A/B testing, hypothesis design, statistical rigorinstrumentation-*β Event tracking, data modeling, collection best practicesdashboard-*β Visualization, stakeholder reporting, self-serve analytics
Core Frameworks
AARRR (Pirate Metrics)
| Stage | Question | Key Metrics |
|---|---|---|
| Acquisition | Where do users come from? | Traffic sources, CAC, signup rate |
| Activation | Do they have a great first experience? | Time-to-value, setup completion, aha moment |
| Retention | Do they come back? | DAU/MAU, D1/D7/D30 retention, churn |
| Revenue | Do they pay? | Conversion rate, ARPU, LTV |
| Referral | Do they tell others? | NPS, referral rate, viral coefficient |
HEART Framework (Google)
| Dimension | Definition | Signal Types |
|---|---|---|
| Happiness | User attitudes, satisfaction | NPS, CSAT, surveys |
| Engagement | Depth of involvement | Sessions, time-in-app, actions/session |
| Adoption | New users/features uptake | New users, feature adoption % |
| Retention | Continued usage over time | Retention curves, churn rate |
| Task Success | Efficiency and completion | Task completion, error rate, time-on-task |
The Metrics Hierarchy
βββββββββββββββββββ
β North Star β β Single metric that matters most
β Metric β
βββββββββββββββββββ€
β Primary β β 3-5 key performance indicators
β KPIs β
βββββββββββββββββββ€
β Supporting β β Diagnostic and health metrics
β Metrics β
βββββββββββββββββββ€
β Operational β β Day-to-day tracking
β Metrics β
βββββββββββββββββββ
Retention Analysis Types
βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ
β RETENTION VIEWS β
βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ€
β N-Day Retention β % who return on exactly day N β
β Unbounded β % who return on or after day N β
β Bracket Retention β % who return within a time window β
β Rolling Retention β % still active after N days β
βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ
Experimentation Rigor Ladder
| Level | Approach | When to Use |
|---|---|---|
| 1. Gut | Ship and hope | Never for important features |
| 2. Qualitative | User research, feedback | Early exploration |
| 3. Observational | Pre/post analysis | Low-risk changes |
| 4. Quasi-experiment | Cohort comparison | When randomization hard |
| 5. A/B Test | Randomized control | Optimization, validation |
| 6. Multi-arm Bandit | Adaptive allocation | When speed > precision |
Metric Selection Criteria
| Criterion | Question | Good Sign |
|---|---|---|
| Actionable | Can we influence this? | Direct lever exists |
| Accessible | Can we measure it reliably? | <5% missing data |
| Auditable | Can we debug anomalies? | Clear calculation logic |
| Aligned | Does it tie to business value? | Executive cares |
| Attributable | Can we trace changes to causes? | A/B testable |
Anti-Patterns
- Vanity metrics β Tracking what looks good, not what drives decisions
- Metric overload β 50 dashboards, zero insights
- Lagging only β Measuring outcomes without predictive indicators
- Silent failures β No alerting on data quality issues
- HiPPO-driven β Highest-paid person's opinion beats data
- P-hacking β Running tests until you get significance
- Ship and forget β Launching features without success criteria
- Segment blindness β Looking only at averages, missing cohort differences
# Supported AI Coding Agents
This skill is compatible with the SKILL.md standard and works with all major AI coding agents:
Learn more about the SKILL.md standard and how to use these skills with your preferred AI coding agent.