xlite-dev / ffpa-attnLinks

📚FFPA(Split-D): Extend FlashAttention with Split-D for large headdim, O(1) GPU SRAM complexity, 1.8x~3x↑🎉 faster than SDPA EA.
184Updated 3 weeks ago

Alternatives and similar repositories for ffpa-attn

Users that are interested in ffpa-attn are comparing it to the libraries listed below

Sorting: