fla-org / flash-linear-attention

πŸš€ Efficient implementations of state-of-the-art linear attention models in Torch and Triton
β˜†1,926Updated this week

Alternatives and similar repositories for flash-linear-attention:

Users that are interested in flash-linear-attention are comparing it to the libraries listed below