Picture for Liu Xiao

Liu Xiao

Cross-attention for State-based model RWKV-7

Add code
Apr 19, 2025
Viaarxiv icon

Millions of States: Designing a Scalable MoE Architecture with RWKV-7 Meta-learner

Add code
Apr 11, 2025
Viaarxiv icon

State Tuning: State-based Test-Time Scaling on RWKV-7

Add code
Apr 07, 2025
Viaarxiv icon

RWKVTTS: Yet another TTS based on RWKV-7

Add code
Apr 04, 2025
Viaarxiv icon

BlackGoose Rimer: Harnessing RWKV-7 as a Simple yet Superior Replacement for Transformers in Large-Scale Time Series Modeling

Add code
Mar 08, 2025
Viaarxiv icon

ARWKV: Pretrain is not what we need, an RNN-Attention-Based Language Model Born from Transformer

Add code
Jan 26, 2025
Viaarxiv icon