xlite-dev / ffpa-attnLinks

⚡️FFPA: Extend FlashAttention-2 with Split-D, achieve ~O(1) SRAM complexity for large headdim, 1.8x~3x↑ vs SDPA.
186Updated last month

Alternatives and similar repositories for ffpa-attn

Users that are interested in ffpa-attn are comparing it to the libraries listed below

Sorting: