Picture for Disen Lan

Disen Lan

Linear-MoE: Linear Sequence Modeling Meets Mixture-of-Experts

Add code
Mar 07, 2025
Viaarxiv icon

Liger: Linearizing Large Language Models to Gated Recurrent Structures

Add code
Mar 03, 2025
Viaarxiv icon

MoM: Linear Sequence Modeling with Mixture-of-Memories

Add code
Feb 19, 2025
Viaarxiv icon

LASP-2: Rethinking Sequence Parallelism for Linear Attention and Its Hybrid

Add code
Feb 11, 2025
Viaarxiv icon

Time-SSM: Simplifying and Unifying State Space Models for Time Series Forecasting

Add code
May 25, 2024
Figure 1 for Time-SSM: Simplifying and Unifying State Space Models for Time Series Forecasting
Figure 2 for Time-SSM: Simplifying and Unifying State Space Models for Time Series Forecasting
Figure 3 for Time-SSM: Simplifying and Unifying State Space Models for Time Series Forecasting
Figure 4 for Time-SSM: Simplifying and Unifying State Space Models for Time Series Forecasting
Viaarxiv icon