zechenzhangAGI

Zechen Zhang

@zechenzhangAGI

Building the future of AI-human collaborations

82 skills 140,384 total stars

find ~/zechenzhangAGI/ -name "*.skill"

Evaluates LLMs across 60+ academic benchmarks (MMLU, HumanEval, GSM8K, TruthfulQA, HellaSwag). Use when benchmarking...

Implements and trains LLMs using Lightning AI's LitGPT with 20+ pretrained architectures (Llama, Gemma, Phi, Qwen,...

Runs LLM inference on CPU, Apple Silicon, and consumer GPUs without NVIDIA hardware. Use for edge deployment,...

Optimizes LLM inference with NVIDIA TensorRT for maximum throughput and lowest latency. Use for production...

Simple Preference Optimization for LLM alignment. Reference-free alternative to DPO with better performance (+6.4...

Track ML experiments, manage model registry with versioning, deploy models to production, and reproduce experiments...

Scalable data processing for ML workloads. Streaming execution across CPU/GPU, supports Parquet/CSV/JSON/images....

Track ML experiments with automatic logging, visualize training in real-time, optimize hyperparameters with sweeps,...

Multi-agent orchestration framework for autonomous AI collaboration. Use when building teams of specialized agents...

Meta's 7-8B specialized moderation model for LLM input/output filtering. 6 safety categories - violence/hate, sexual...

Data framework for building LLM applications with RAG. Specializes in document ingestion (300+ connectors),...

Provides guidance for PyTorch-native agentic RL using torchforge, Meta's library separating infra from algorithms....

Simplest distributed training API. 4 lines to add distributed support to any PyTorch script. Unified API for...

Write publication-ready ML/AI papers for NeurIPS, ICML, ICLR, ACL, AAAI, COLM. Use when drafting papers from...

State-space model with O(n) complexity vs Transformers' O(n²). 5× faster inference, million-token sequences, no KV...

Fast tokenizers optimized for research and production. Rust-based implementation tokenizes 1GB in <20 seconds....

Fine-tune LLMs using reinforcement learning with TRL - SFT for instruction tuning, DPO for preference alignment,...

Expert guidance for distributed training with DeepSpeed - ZeRO optimization stages, pipeline parallelism,...