Picture for Pinxue Zhao

Pinxue Zhao

HMoE: Heterogeneous Mixture of Experts for Language Modeling

Add code
Aug 20, 2024
Viaarxiv icon

Efficiently Training 7B LLM with 1 Million Sequence Length on 8 GPUs

Add code
Jul 16, 2024
Viaarxiv icon