xlite-dev / ffpa-attnView on GitHub
FFPA: Extend FlashAttention-2 with Split-D, ~O(1) SRAM complexity for large headdim, 1.8x~3x↑🎉 vs SDPA.
276Apr 29, 2026Updated this week

Alternatives and similar repositories for ffpa-attn

Users that are interested in ffpa-attn are comparing it to the libraries listed below. We may earn a commission when you buy through links labeled 'Ad' on this page.

Sorting:

Are these results useful?