BlackGoose Rimer: Harnessing RWKV-7 as a Simple yet Superior Replacement for Transformers in Large-Scale Time Series Modeling Paper β’ 2503.06121 β’ Published 6 days ago β’ 5
BlackGoose Rimer: Harnessing RWKV-7 as a Simple yet Superior Replacement for Transformers in Large-Scale Time Series Modeling Paper β’ 2503.06121 β’ Published 6 days ago β’ 5 β’ 2
ARWKV: Pretrain is not what we need, an RNN-Attention-Based Language Model Born from Transformer Paper β’ 2501.15570 β’ Published Jan 26 β’ 23
ARWKV: Pretrain is not what we need, an RNN-Attention-Based Language Model Born from Transformer Paper β’ 2501.15570 β’ Published Jan 26 β’ 23 β’ 2
recursal/QRWKV6-32B-Instruct-Preview-v0.1 Text Generation β’ Updated about 6 hours ago β’ 787 β’ 71