Researcher, this channel is just my reading list.
Currently, my research focuses on RNNs in large language models (LLMs). I’m interested in exploring how to adapt transformer-based models, like the 671B R1, to use RNN attention. You can check out my ongoing work on ARWKV here.
huggingface.co/papers/2501.15570

X: x.com/xiaolGo
Papers: scholar.google.com/citations?user=TPJYxnkAAAAJ



26:17

Shared 1 day ago

66 views

25:44

Shared 2 weeks ago

187 views

7:10

Shared 1 month ago

894 views

7:10

Shared 1 month ago

62 views

6:35

Shared 1 month ago

60 views