shreyansh26 / FlashAttention-PyTorchLinks
Implementation of FlashAttention in PyTorch
☆155Updated 6 months ago
Alternatives and similar repositories for FlashAttention-PyTorch
Users that are interested in FlashAttention-PyTorch are comparing it to the libraries listed below
Sorting:
- ☆140Updated last week
- TransMLA: Multi-Head Latent Attention Is All You Need☆327Updated last week
- DeepSeek Native Sparse Attention pytorch implementation☆73Updated 4 months ago
- Get down and dirty with FlashAttention2.0 in pytorch, plug in and play no complex CUDA kernels☆105Updated last year
- [ICLR 2025] COAT: Compressing Optimizer States and Activation for Memory-Efficient FP8 Training☆215Updated last month
- ☆116Updated last month
- ☆145Updated 4 months ago
- Lightning Attention-2: A Free Lunch for Handling Unlimited Sequence Lengths in Large Language Models