Experience
Architecting scalable infrastructure and models for humanized machine intelligence. (Hiring! Reach out if interested.)
Optimized training infrastructure for LLM & Video/Audio workloads.
Projects: Step-1o, Step-1.5V, Step-2, Step-3 AFD, StepMesh.
Pioneered the engineering frameworks for LLM pretraining and RL in Seed.
Prior projects include: BytePS/ByteCCL; Sparse MoE modeling and training.
(1) Experienced in architecting, optimizing and diagnosing large scale systems over 10,000 GPUs.
(2) Responsible for training several (multi-modal) LLMs from scratch (cumulative FLOPs > 1e26).
Selected Publications
See a full list at Google Scholar.