Picture for Weigao Sun

Weigao Sun

A Survey of Efficient Reasoning for Large Reasoning Models: Language, Multimodality, and Beyond

Add code
Mar 27, 2025
Viaarxiv icon

Linear-MoE: Linear Sequence Modeling Meets Mixture-of-Experts

Add code
Mar 07, 2025
Viaarxiv icon

Liger: Linearizing Large Language Models to Gated Recurrent Structures

Add code
Mar 03, 2025
Viaarxiv icon

MoM: Linear Sequence Modeling with Mixture-of-Memories

Add code
Feb 19, 2025
Viaarxiv icon

LASP-2: Rethinking Sequence Parallelism for Linear Attention and Its Hybrid

Add code
Feb 11, 2025
Viaarxiv icon

MiniMax-01: Scaling Foundation Models with Lightning Attention

Add code
Jan 14, 2025
Viaarxiv icon

LLaMA-MoE v2: Exploring Sparsity of LLaMA from Perspective of Mixture-of-Experts with Post-Training

Add code
Nov 24, 2024
Viaarxiv icon

Scaling Laws for Linear Complexity Language Models

Add code
Jun 24, 2024
Viaarxiv icon

Unlocking the Secrets of Linear Complexity Sequence Model from A Unified Perspective

Add code
May 27, 2024
Viaarxiv icon

Various Lengths, Constant Speed: Efficient Language Modeling with Lightning Attention

Add code
May 27, 2024
Viaarxiv icon