Picture for Lin Yueyu

Lin Yueyu

Cross-attention for State-based model RWKV-7

Add code
Apr 19, 2025
Viaarxiv icon

Millions of States: Designing a Scalable MoE Architecture with RWKV-7 Meta-learner

Add code
Apr 11, 2025
Viaarxiv icon

State Tuning: State-based Test-Time Scaling on RWKV-7

Add code
Apr 07, 2025
Viaarxiv icon

ARWKV: Pretrain is not what we need, an RNN-Attention-Based Language Model Born from Transformer

Add code
Jan 26, 2025
Viaarxiv icon