zechenzhangAGI

Zechen Zhang

@zechenzhangAGI

Building the future of AI-human collaborations

82 skills 140,384 total stars

find ~/zechenzhangAGI/ -name "*.skill"

Provides guidance for training LLMs with reinforcement learning using verl (Volcano Engine RL). Use when...

LLM observability platform for tracing, evaluation, and monitoring. Use when debugging LLM applications, evaluating...

High-level PyTorch framework with Trainer class, automatic distributed training (DDP/FSDP/DeepSpeed), callbacks...

PyTorch library for audio generation including text-to-music (MusicGen) and text-to-sound (AudioGen). Use when you...

Language-independent tokenizer treating text as raw Unicode. Supports BPE and Unigram algorithms. Fast (50k...

OpenAI's model connecting vision and language. Enables zero-shot image classification, image-text matching, and...

Reserved and on-demand GPU cloud instances for ML training and inference. Use when you need dedicated GPU instances...

Foundation model for image segmentation with zero-shot transfer. Use when you need to segment any object in images...

Expert guidance for fine-tuning LLMs with LLaMA-Factory - WebUI no-code, 100+ models, 2/3/4/5/6/8-bit QLoRA,...

OpenAI's general-purpose speech recognition model. Supports 99 languages, transcription, translation to English, and...

Extend context windows of transformer models using RoPE, YaRN, ALiBi, and position interpolation techniques. Use...

Educational GPT implementation in ~300 lines. Reproduces GPT-2 (124M) on OpenWebText. Clean, hackable code for...

Quantizes LLMs to 8-bit or 4-bit for 50-75% memory reduction with minimal accuracy loss. Use when GPU memory is...

Write publication-ready ML/AI papers for NeurIPS, ICML, ICLR, ACL, AAAI, COLM. Use when drafting papers from...

State-space model with O(n) complexity vs Transformers' O(n²). 5× faster inference, million-token sequences, no KV...

Fast tokenizers optimized for research and production. Rust-based implementation tokenizes 1GB in <20 seconds....

Fine-tune LLMs using reinforcement learning with TRL - SFT for instruction tuning, DPO for preference alignment,...

Expert guidance for distributed training with DeepSpeed - ZeRO optimization stages, pipeline parallelism,...

Framework for state-of-the-art sentence, text, and image embeddings. Provides 5000+ pre-trained models for semantic...