howard-hou / RWKV-XLinks
RWKV-X is a Linear Complexity Hybrid Language Model based on the RWKV architecture, integrating Sparse Attention to improve the model's long sequence processing capabilities.
☆53Updated 5 months ago
Alternatives and similar repositories for RWKV-X
Users that are interested in RWKV-X are comparing it to the libraries listed below
Sorting:
- Here we will test various linear attention designs.☆62Updated last year
- RADLADS training code☆35Updated 7 months ago
- Official Repo for Error-Free Linear Attention is a Free Lunch: Exact Solution from Continuous-Time Dynamics