TorchRWKV / flash-linear-attention

Efficient implementations of state-of-the-art linear attention models in Pytorch and Triton
13Updated this week

Related projects

Alternatives and complementary repositories for flash-linear-attention