rlin27 / DeButLinks
Codes of the paper Deformable Butterfly: A Highly Structured and Sparse Linear Transform.
☆13Updated 3 years ago
Alternatives and similar repositories for DeBut
Users that are interested in DeBut are comparing it to the libraries listed below
Sorting:
- Implementation of a Transformer using ReLA (Rectified Linear Attention) from https://arxiv.org/abs/2104.07012☆49Updated 3 years ago
- ☆14Updated 2 years ago
- The official repository for our paper "The Dual Form of Neural Networks Revisited: Connecting Test Time Predictions to Training Patterns …☆16Updated 3 months ago
- Xmixers: A collection of SOTA efficient token/channel mixers☆14Updated last week
- sigma-MoE layer☆20Updated last year
- ☆21Updated 2 years ago
- [NeurIPS 2023] Sparse Modular Activation for Efficient Sequence Modeling☆40Updated last year
- Implementation for ACProp ( Momentum centering and asynchronous update for adaptive gradient methdos, NeurIPS 2021)☆16Updated 3 years ago
- ☆29Updated 2 years ago
- DropIT: Dropping Intermediate Tensors for Memory-Efficient DNN Training (ICLR 2023)