Dao-AILab / flash-attention

Fast and memory-efficient exact attention
β˜†13,401Updated this week

Related projects: β“˜