Use when executing implementation plans with independent tasks in the current session
Use when executing implementation plans with independent tasks in the current session
Use when executing implementation plans with independent tasks in the current session
Use when executing implementation plans with independent tasks in the current session
Use when executing implementation plans with independent tasks in the current session
Use when executing implementation plans with independent tasks in the current session
Use when executing implementation plans with independent tasks in the current session
Use when executing implementation plans with independent tasks in the current session
Use when executing implementation plans with independent tasks in the current session
Use when executing implementation plans with independent tasks in the current session
Use when executing implementation plans with independent tasks in the current session
Use when executing implementation plans with independent tasks in the current session
Use when executing implementation plans with independent tasks in the current session
Comprehensive guide for using Claude Agent SDK to build AI agents that can read files, execute commands, edit code, and perform complex workflows. Use when (1) Building autonomous AI agents with...
Plan and execute launch marketing by producing a Launch Marketing Pack (launch brief, hook/sizzle, channel plan, PR outreach kit, internal readiness kit, execution checklist, measurement +...
Expert guidance for Fully Sharded Data Parallel training with PyTorch FSDP - parameter sharding, mixed precision, CPU offloading, FSDP2
Expert guidance for Fully Sharded Data Parallel training with PyTorch FSDP - parameter sharding, mixed precision, CPU offloading, FSDP2
Drive development using delegated agent workflows. Coordinates multi-agent task execution with proper supervision and result integration.
Expert guidance for distributed training with DeepSpeed - ZeRO optimization stages, pipeline parallelism, FP16/BF16/FP8, 1-bit Adam, sparse attention
Expert guidance for distributed training with DeepSpeed - ZeRO optimization stages, pipeline parallelism, FP16/BF16/FP8, 1-bit Adam, sparse attention