Picture for Weigao Sun

Weigao Sun

A Survey of Efficient Reasoning for Large Reasoning Models: Language, Multimodality, and Beyond

Add code
Mar 27, 2025
Viaarxiv icon

Linear-MoE: Linear Sequence Modeling Meets Mixture-of-Experts

Add code
Mar 07, 2025
Viaarxiv icon

Liger: Linearizing Large Language Models to Gated Recurrent Structures

Add code
Mar 03, 2025
Viaarxiv icon

MoM: Linear Sequence Modeling with Mixture-of-Memories

Add code
Feb 19, 2025
Viaarxiv icon

LASP-2: Rethinking Sequence Parallelism for Linear Attention and Its Hybrid

Add code
Feb 11, 2025
Viaarxiv icon

MiniMax-01: Scaling Foundation Models with Lightning Attention

Add code
Jan 14, 2025
Viaarxiv icon

LLaMA-MoE v2: Exploring Sparsity of LLaMA from Perspective of Mixture-of-Experts with Post-Training

Add code
Nov 24, 2024
Viaarxiv icon

Scaling Laws for Linear Complexity Language Models

Add code
Jun 24, 2024
Viaarxiv icon

Various Lengths, Constant Speed: Efficient Language Modeling with Lightning Attention

Add code
May 27, 2024
Viaarxiv icon

Unlocking the Secrets of Linear Complexity Sequence Model from A Unified Perspective

Add code
May 27, 2024
Viaarxiv icon