00ffcc / chunkRWKV6Links
continous batching and parallel acceleration for RWKV6
☆22Updated last year
Alternatives and similar repositories for chunkRWKV6
Users that are interested in chunkRWKV6 are comparing it to the libraries listed below
Sorting:
- Flash-Linear-Attention models beyond language☆20Updated 3 months ago
- ☆22Updated last year
- ☆32Updated last year
- ☆27Updated 4 months ago
- [NeurIPS 2023 spotlight] Official implementation of HGRN in our NeurIPS 2023 paper - Hierarchically Gated Recurrent Neural Network for Se…☆66Updated last year
- Flash Attention in 300-500 lines of CUDA/C++☆36Updated 3 months ago
- Fast and memory-efficient exact attention☆74Updated 9 months ago
- Xmixers: A collection of SOTA efficient token/channel mixers☆29Updated 3 months ago
- ☆57Updated last year