lucidrains / performer-pytorchView external linksLinks
An implementation of Performer, a linear attention-based transformer, in Pytorch
☆1,172Feb 2, 2022Updated 4 years ago
Alternatives and similar repositories for performer-pytorch
Users that are interested in performer-pytorch are comparing it to the libraries listed below
Sorting:
- Pytorch library for fast transformer implementations☆1,761Mar 23, 2023Updated 2 years ago
- Reformer, the efficient Transformer, in Pytorch☆2,193Jun 21, 2023Updated 2 years ago
- Transformer based on a variant of attention that is linear complexity in respect to sequence length☆827May 5, 2024Updated last year
- Implementation of Linformer for Pytorch☆305Jan 5, 2024Updated 2 years ago
- My take on a practical implementation of Linformer for Pytorch.☆422Jul 27, 2022Updated 3 years ago
- Long Range Arena for Benchmarking Efficient Transformers☆777Dec 16, 2023Updated 2 years ago
- A concise but complete full-attention transformer with a set of promising experimental features from various papers☆5,800Feb 7, 2026Updated last week
- Fully featured implementation of Routing Transformer☆300Nov 6, 2021Updated 4 years ago
- ☆388Oct 18, 2023Updated 2 years ago
- Implementation of Long-Short Transformer, combining local and global inductive biases for attention over long sequences, in Pytorch☆120Aug 4, 2021Updated 4 years ago
- DeLighT: Very Deep and Light-Weight Transformers☆469Oct 16, 2020Updated 5 years ago
- [ICLR 2020] Lite Transformer with Long-Short Range Attention☆611Jul 11, 2024Updated last year
- Implementation of Feedback Transformer in Pytorch☆108Mar 2, 2021Updated 4 years ago
- Flexible and powerful tensor operations for readable and reliable code (for pytorch, jax, TF and others)☆9,395Jan 26, 2026Updated 3 weeks ago
- Longformer: The Long-Document Transformer☆2,186Feb 8, 2023Updated 3 years ago
- FastFormers - highly efficient transformer models for NLU☆709Mar 21, 2025Updated 10 months ago
- Code for the Shortformer model, from the ACL 2021 paper by Ofir Press, Noah A. Smith and Mike Lewis.☆147Jul 26, 2021Updated 4 years ago
- Pytorch implementation of Compressive Transformers, from Deepmind☆163Oct 4, 2021Updated 4 years ago
- Implementation of Bottleneck Transformer in Pytorch☆677Sep 20, 2021Updated 4 years ago
- Implementation of TransGanFormer, an all-attention GAN that combines the finding from the recent GanFormer and TransGan paper☆155Apr 27, 2021Updated 4 years ago
- Hopfield Networks is All You Need☆1,897Apr 23, 2023Updated 2 years ago
- Sinkhorn Transformer - Practical implementation of Sparse Sinkhorn Attention☆270Aug 10, 2021Updated 4 years ago
- Repository for NeurIPS 2020 Spotlight "AdaBelief Optimizer: Adapting stepsizes by the belief in observed gradients"☆1,068Aug 9, 2024Updated last year
- ☆221Jun 8, 2020Updated 5 years ago
- Implementation / replication of DALL-E, OpenAI's Text to Image Transformer, in Pytorch☆5,629Feb 17, 2024Updated last year
- Implementation of Multistream Transformers in Pytorch☆54Jul 31, 2021Updated 4 years ago
- Implementation of the convolutional module from the Conformer paper, for use in Transformers☆433May 17, 2023Updated 2 years ago
- Google Research☆37,261Updated this week
- Structured state space sequence models☆2,842Jul 17, 2024Updated last year
- torch-optimizer -- collection of optimizers for Pytorch☆3,161Mar 22, 2024Updated last year
- Pretrain, finetune ANY AI model of ANY size on 1 or 10,000+ GPUs with zero code changes.☆30,823Feb 4, 2026Updated last week
- SentAugment is a data augmentation technique for NLP that retrieves similar sentences from a large bank of sentences. It can be used in c…☆359Feb 22, 2022Updated 3 years ago
- An implementation of the efficient attention module.☆328Nov 30, 2020Updated 5 years ago
- [NeurIPS‘2021] "TransGAN: Two Pure Transformers Can Make One Strong GAN, and That Can Scale Up", Yifan Jiang, Shiyu Chang, Zhangyang Wang☆1,690Nov 3, 2022Updated 3 years ago
- Facebook AI Research Sequence-to-Sequence Toolkit written in Python.☆32,153Sep 30, 2025Updated 4 months ago
- Axial Positional Embedding for Pytorch☆84Feb 25, 2025Updated 11 months ago
- Official DeiT repository☆4,323Mar 15, 2024Updated last year
- Implementation of SE3-Transformers for Equivariant Self-Attention, in Pytorch. This specific repository is geared towards integration wit…☆325Aug 28, 2025Updated 5 months ago
- Official code Cross-Covariance Image Transformer (XCiT)☆674Sep 28, 2021Updated 4 years ago