Researcher, this channel is just my reading list.
Currently, my research focuses on RNNs in large language models (LLMs). I’m interested in exploring how to adapt transformer-based models, like the 671B R1, to use RNN attention. You can check out my ongoing work on ARWKV here.
huggingface.co/papers/2501.15570

X: x.com/xiaolGo
Papers: scholar.google.com/citations?user=TPJYxnkAAAAJ



Xiaol.x

This is how i read 600 papers in half a year.

5 months ago | [YT] | 9

Xiaol.x

cohere.com/blog/embed-4

good for rag or search

6 months ago | [YT] | 2

Xiaol.x

Top 30 most popular arXiv papers in the last 30 days in hacker news
from @susumuota

6 months ago | [YT] | 20

Xiaol.x

1 year ago | [YT] | 0