404kidwiz

performance-testing

6
0
# Install this skill:
npx skills add 404kidwiz/claude-supercode-skills --skill "performance-testing"

Install specific skill from multi-skill repository

# Description

Performance testing specialist for load testing, stress testing, and performance optimization across applications and infrastructure

# SKILL.md


name: performance-testing
description: Performance testing specialist for load testing, stress testing, and performance optimization across applications and infrastructure


Performance Testing Skill

Purpose

Provides comprehensive performance testing expertise specializing in load testing, stress testing, and endurance testing for applications, APIs, and infrastructure systems. Validates system behavior under various load conditions and identifies capacity limits.

When to Use

  • Conducting load testing to simulate concurrent users
  • Performing stress testing to find breaking points
  • Running endurance tests to detect memory leaks
  • Validating system scalability under increased load
  • Measuring response times and throughput metrics
  • Analyzing resource utilization patterns

Examples

Example 1: E-commerce Platform Load Testing

Scenario: An e-commerce platform preparing for Black Friday needs to validate performance under 10x expected traffic.

Implementation:
1. Analyzed historical traffic patterns to model realistic load
2. Created JMeter test scenarios for critical paths (browse, cart, checkout)
3. Simulated 10,000 concurrent users with ramp-up period
4. Tested database queries under load, identified N+1 issues
5. Validated auto-scaling triggers and thresholds

Results:
- Identified 3 critical bottlenecks (2 DB, 1 API)
- Reduced average response time from 2.3s to 450ms
- Validated system handles 50,000 concurrent users
- Confirmed auto-scaling activates within 45 seconds

Example 2: API Performance Benchmarking

Scenario: A financial services API needs performance validation against SLA requirements (<200ms P99).

Implementation:
1. Created k6 test scripts for all API endpoints
2. Tested with various concurrency levels (100, 500, 1000)
3. Analyzed response time distributions (P50, P90, P95, P99)
4. Profiled database queries causing slow responses
5. Implemented query optimizations and caching

Results:
- P99 latency reduced from 850ms to 145ms
- Throughput increased from 500 to 2,500 RPS
- All endpoints meet SLA requirements
- Created performance regression test suite

Example 3: Microservices Chaos Testing

Scenario: A microservices architecture needs validation under partial service failures.

Implementation:
1. Designed chaos test scenarios for service failures
2. Implemented latency injection and error responses
3. Tested circuit breaker functionality and recovery
4. Validated graceful degradation behaviors
5. Measured end-to-end request flows under chaos

Results:
- Identified 2 services without proper circuit breakers
- Validated recovery times under various failure scenarios
- Confirmed system degrades gracefully (not catastrophically)
- Documented recovery procedures for each failure mode

Best Practices

Test Design

  • Realistic Workload Modeling: Base load patterns on production traffic analysis
  • Think Time Inclusion: Include realistic user pause times between requests
  • Data Parameterization: Use varied test data to avoid caching artifacts
  • Comprehensive Coverage: Test all critical user journeys, not just happy paths

Execution

  • Production-Like Environments: Test in environments matching production configuration
  • Proper Warm-up: Include ramp-up periods before measurement
  • Extended Duration: Run stress tests long enough to discover issues
  • Error Rate Monitoring: Track both response times and error rates

Analysis

  • Percentile Analysis: Report P50, P90, P95, P99, not just averages
  • Baseline Comparison: Always compare against established baselines
  • Trend Tracking: Monitor performance over time, not just snapshots
  • Focused Metrics: Track SLA-relevant metrics, avoid metric overload

Tooling

  • Test Maintenance: Treat tests as production code, maintain and update
  • CI/CD Integration: Automate performance tests in deployment pipelines
  • Adequate Resources: Ensure load generators can generate required load
  • Result Analysis: Implement automated analysis and alerting

Overview

Performance testing expert specializing in load testing, stress testing, and performance optimization for applications, APIs, and infrastructure systems.

Performance Testing Types

Load Testing

  • Concurrent user simulation
  • Transaction volume testing
  • Scalability assessment
  • Resource utilization analysis
  • Response time measurement

Stress Testing

  • Breaking point identification
  • Failure mode analysis
  • Recovery time measurement
  • Resource exhaustion testing
  • System stability validation

Endurance Testing

  • Long-term stability assessment
  • Memory leak detection
  • Performance degradation analysis
  • Resource growth monitoring
  • System sustainability testing

Performance Testing Tools

Open Source Tools

  • Apache JMeter - Comprehensive performance testing
  • Gatling - High-performance load testing
  • k6 - Modern load testing with JavaScript
  • Locust - Python-based load testing
  • WRK - HTTP benchmarking tool

Commercial Solutions

  • LoadRunner Professional
  • NeoLoad
  • Silk Performer
  • BlazeMeter
  • WebLOAD

Cloud-Based Platforms

  • AWS Load Testing
  • Azure Load Testing
  • Google Cloud Load Testing
  • k6 Cloud
  • BlazeMeter Cloud

Performance Metrics & Analysis

Key Performance Indicators

# Example patterns for performance analysis
grep -r "response_time\|latency\|throughput" logs/ --include="*.log" --include="*.txt"
grep -r "cpu\|memory\|disk" monitoring/ --include="*.metrics" --include="*.json"
grep -r "concurrent\|connections\|requests" load_tests/ --include="*.js" --include="*.py"

Response Time Analysis

  • Average response time
  • Median (50th percentile)
  • 90th, 95th, 99th percentile analysis
  • Maximum response time
  • Response time distribution

Throughput Metrics

  • Requests per second (RPS)
  • Transactions per second (TPS)
  • Data transfer rates
  • Concurrent user capacity
  • Peak load handling

Resource Utilization

  • CPU usage monitoring
  • Memory consumption tracking
  • Disk I/O analysis
  • Network bandwidth usage
  • Database connection pooling

Test Design & Execution

Test Scenario Planning

  • User journey mapping
  • Business process modeling
  • Peak load simulation
  • Ramp-up strategies
  • Think time implementation

Load Profile Design

  • Constant load patterns
  • Spike testing scenarios
  • Gradual ramp-up loads
  • Custom load curves
  • Real-world traffic simulation

Test Data Management

  • Test data generation
  • Parameterization strategies
  • Data variety creation
  • Database state management
  • Privacy protection measures

Application-Specific Testing

Web Application Performance

  • Page load time analysis
  • Asset loading optimization
  • JavaScript execution performance
  • CSS rendering performance
  • Third-party dependency impact

API Performance Testing

  • RESTful API testing
  • GraphQL performance
  • SOAP web service testing
  • Authentication overhead
  • Rate limiting validation

Database Performance

  • Query optimization
  • Index efficiency analysis
  • Connection pooling
  • Database scaling
  • Lock contention analysis

Mobile Application Testing

  • Network condition simulation
  • Device performance variability
  • Battery consumption analysis
  • App startup time
  • Memory usage patterns

Advanced Performance Testing

Distributed Testing

  • Multiple load generators
  • Geographic distribution
  • Network latency simulation
  • Bandwidth throttling
  • Cloud-based load generation

Real User Monitoring (RUM)

  • Front-end performance tracking
  • User experience metrics
  • Geographic performance analysis
  • Device-specific performance
  • Browser compatibility impact

Continuous Performance Testing

  • Integration with CI/CD
  • Automated regression testing
  • Performance threshold validation
  • Alerting and notification
  • Trend analysis and reporting

Performance Analysis & Optimization

Bottleneck Identification

  • CPU-bound analysis
  • Memory optimization
  • I/O bottleneck detection
  • Network latency analysis
  • Database query optimization

Profiling & Diagnostics

  • Application profiling
  • System call analysis
  • Memory leak detection
  • Thread contention analysis
  • Garbage collection tuning

Caching Strategies

  • Application-level caching
  • Database query caching
  • Content Delivery Networks
  • Browser caching optimization
  • Distributed cache implementation

Monitoring & Observability

Application Performance Monitoring (APM)

  • Real-time performance tracking
  • Distributed tracing
  • Error rate monitoring
  • Custom metrics collection
  • Performance dashboards

Infrastructure Monitoring

  • Server resource monitoring
  • Network performance tracking
  • Database performance metrics
  • Cloud resource utilization
  • Container performance analysis

Log Analysis

  • Performance-related log patterns
  • Error log correlation
  • Access log analysis
  • Custom performance logging
  • Log aggregation and search

Performance Testing Automation

Test Automation Frameworks

  • JMeter automation
  • Gatling scripting
  • k6 JavaScript automation
  • Python-based automation
  • CI/CD integration

Continuous Integration

  • Automated test execution
  • Performance regression detection
  • Automated reporting
  • Threshold validation
  • Failure notification systems

Cloud-Based Automation

  • Scalable load generation
  • Geographic distribution
  • On-demand resource provisioning
  • Cost optimization
  • Multi-cloud strategies

Performance Testing in Different Environments

Development Environment

  • Early performance validation
  • Unit-level performance testing
  • Local benchmarking
  • Development feedback loops
  • Performance best practices

Staging Environment

  • Production-like testing
  • Capacity planning validation
  • Performance regression testing
  • Integration performance testing
  • Pre-deployment validation

Production Monitoring

  • Real-time performance tracking
  • Performance SLA monitoring
  • User experience measurement
  • Incident response
  • Performance optimization cycles

Reporting & Documentation

Performance Test Reports

  • Executive summary
  • Detailed test results
  • Performance comparisons
  • Bottleneck analysis
  • Optimization recommendations

Performance Dashboards

  • Real-time metrics display
  • Historical trend analysis
  • SLA compliance tracking
  • Resource utilization charts
  • User experience metrics

Benchmarking Documentation

  • Baseline performance metrics
  • Industry comparisons
  • Competitive analysis
  • Performance goals setting
  • Progress tracking

Specific Industry Expertise

E-commerce Performance

  • Shopping cart performance
  • Checkout process optimization
  • Search functionality testing
  • Product catalog performance
  • Payment processing optimization

Financial Services

  • Trading system performance
  • Risk calculation speed
  • Report generation performance
  • Data processing efficiency
  • Regulatory compliance requirements

Healthcare Systems

  • Patient data retrieval
  • Medical imaging performance
  • Real-time monitoring systems
  • Data privacy compliance
  • System availability requirements

Deliverables

Test Plans & Scenarios

  • Comprehensive test strategies
  • Detailed test scenarios
  • Load profile specifications
  • Test data requirements
  • Execution schedules

Performance Reports

  • Detailed analysis reports
  • Executive summaries
  • Technical recommendations
  • Optimization roadmaps
  • Performance benchmarks

Automation Frameworks

  • Custom testing scripts
  • CI/CD integration code
  • Monitoring setup configurations
  • Alerting system setup
  • Documentation and training materials

Anti-Patterns

Test Design Anti-Patterns

  • Unrealistic Workloads: Tests that don't reflect real usage patterns - profile production traffic first
  • Missing Think Time: Continuous requests without user pause times - include realistic user delays
  • Static Data Only: Tests with no data variation - use parameterized and varied test data
  • Single Scenario Focus: Testing one path only - cover all critical user journeys

Execution Anti-Patterns

  • Test Environment Gap: Testing in non-representative environments - match production configuration
  • No Warm-up: Starting tests without system warm-up - include ramp-up periods
  • Stopping Too Early: Ending tests before finding limits - continue until failure
  • Ignoring Error Rates: Focusing only on response times - monitor error rates too

Analysis Anti-Patterns

  • Averages Only: Relying only on averages - analyze percentiles and distributions
  • No Baselines: Testing without baseline comparisons - establish performance baselines
  • Snapshot Testing: One-time tests without trend tracking - monitor over time
  • Metric Overload: Tracking too many irrelevant metrics - focus on SLA-relevant metrics

Tooling Anti-Patterns

  • Scripted Once: Tests not maintained or updated - treat tests as production code
  • Manual Test Creation: No automation of test generation - generate tests from specs
  • No CI Integration: Performance tests run manually - integrate into CI/CD pipeline
  • Resource Contention: Load generators underpowered - ensure adequate load generation capacity

# Supported AI Coding Agents

This skill is compatible with the SKILL.md standard and works with all major AI coding agents:

Learn more about the SKILL.md standard and how to use these skills with your preferred AI coding agent.