sdbds / SageAttention-for-windowsView on GitHub
Quantized Attention that achieves speedups of 2.1-3.1x and 2.7-5.1x compared to FlashAttention2 and xformers, respectively, without lossing end-to-end metrics across various models.
128Mar 24, 2026Updated this week

Alternatives and similar repositories for SageAttention-for-windows

Users that are interested in SageAttention-for-windows are comparing it to the libraries listed below. We may earn a commission when you buy through links labeled 'Ad' on this page.

Sorting:

Are these results useful?