[ICLR 2020] Lite Transformer with Long-Short Range Attention
☆610Jul 11, 2024Updated last year
Alternatives and similar repositories for lite-transformer
Users that are interested in lite-transformer are comparing it to the libraries listed below. We may earn a commission when you buy through links labeled 'Ad' on this page.
Sorting:
- [ACL'20] HAT: Hardware-Aware Transformers for Efficient Natural Language Processing☆336Jul 14, 2024Updated last year
- DeLighT: Very Deep and Light-Weight Transformers☆469Oct 16, 2020Updated 5 years ago
- Transformer training code for sequential tasks☆609Sep 14, 2021Updated 4 years ago
- Reformer, the efficient Transformer, in Pytorch☆2,192Jun 21, 2023Updated 2 years ago
- ☆221Jun 8, 2020Updated 5 years ago
- Proton VPN Special Offer - Get 70% off • AdSpecial partner offer. Trusted by over 100 million users worldwide. Tested, Approved and Recommended by Experts.
- Pytorch library for fast transformer implementations☆1,765Mar 23, 2023Updated 3 years ago
- a fast and user-friendly runtime for transformer inference (Bert, Albert, GPT2, Decoders, etc) on CPU and GPU.☆1,542Jul 18, 2025Updated 8 months ago
- Longformer: The Long-Document Transformer☆2,189Feb 8, 2023Updated 3 years ago
- FastFormers - highly efficient transformer models for NLU☆709Mar 21, 2025Updated last year
- [JMLR'20] NeurIPS 2019 MicroNet Challenge Efficient Language Modeling, Champion☆42Feb 26, 2021Updated 5 years ago
- Official Pytorch Implementation of Length-Adaptive Transformer (ACL 2021)☆102Nov 2, 2020Updated 5 years ago
- Understanding the Difficulty of Training Transformers☆332May 31, 2022Updated 3 years ago
- [ICLR 2020] Once for All: Train One Network and Specialize it for Efficient Deployment☆1,944Dec 14, 2023Updated 2 years ago
- ☆255Oct 4, 2022Updated 3 years ago
- 1-Click AI Models by DigitalOcean Gradient • AdDeploy popular AI models on DigitalOcean Gradient GPU virtual machines with just a single click and start building anything your business needs.
- ELECTRA: Pre-training Text Encoders as Discriminators Rather Than Generators☆2,370Mar 23, 2024Updated 2 years ago
- SentAugment is a data augmentation technique for NLP that retrieves similar sentences from a large bank of sentences. It can be used in c…