Picture for Liu Xiao

Liu Xiao

BlackGoose Rimer: Harnessing RWKV-7 as a Simple yet Superior Replacement for Transformers in Large-Scale Time Series Modeling

Add code
Mar 08, 2025
Viaarxiv icon

ARWKV: Pretrain is not what we need, an RNN-Attention-Based Language Model Born from Transformer

Add code
Jan 26, 2025
Viaarxiv icon