vllm-project / flash-attention

Fast and memory-efficient exact attention
28Updated 2 weeks ago

Related projects

Alternatives and complementary repositories for flash-attention