Andrew-Tierno / QuantizedTransformer
Implementation of a Quantized Transformer Model
☆17Updated 5 years ago
Related projects: ⓘ
- Implementation of ICLR 2018 paper "Loss-aware Weight Quantization of Deep Networks"☆26Updated 4 years ago
- Deep Neural Network Compression based on Student-Teacher Network☆14Updated last year
- Implementation of NeurIPS 2019 paper "Normalization Helps Training of Quantized LSTM"☆30Updated last month
- ☆26Updated 4 years ago
- Code for paper "Continual and Multi-Task Architecture Search (ACL 2019)"☆41Updated 5 years ago
- caffe implementation of single level quantization☆19Updated 5 years ago
- Revisiting Parameter Sharing for Automatic Neural Channel Number Search, NeurIPS 2020☆20Updated 3 years ago
- Code for paper 'Minimizing FLOPs to Learn Efficient Sparse Representations' published at ICLR 2020☆21Updated 4 years ago
- Codes for accepted paper "Cooperative Pruning in Cross-Domain Deep Neural Network Compression" in IJCAI 2019.☆11Updated 5 years ago
- Implementation of ICLR 2017 paper "Loss-aware Binarization of Deep Networks"☆18Updated 5 years ago
- ☆15Updated 4 years ago
- Codes for AAAI2019 paper: Deep Neural Network Quantization via Layer-Wise Optimization using Limited Training Data☆42Updated 5 years ago
- ☆13Updated 5 years ago
- ☆10Updated this week
- Open Source Neural Machine Translation in PyTorch☆17Updated 5 years ago
- Method to improve inference time for BERT. This is an implementation of the paper titled "PoWER-BERT: Accelerating BERT Inference via Pro…☆58Updated last year
- Block Sparse movement pruning☆77Updated 3 years ago
- The official implementation of You Only Compress Once: Towards Effective and Elastic BERT Compression via Exploit-Explore Stochastic Natu…☆48Updated 3 years ago
- ICML2019 Accepted Paper. Overcoming Multi-Model Forgetting☆13Updated 5 years ago
- PyTorch code for full quantization of DNN using BCGD☆14Updated 5 years ago
- Source code for NAACL 2021 paper "TR-BERT: Dynamic Token Reduction for Accelerating BERT Inference"☆44Updated 2 years ago
- A Unified, Systematic Framework of Structured Weight Pruning for DNNs☆21Updated 6 years ago
- A fully differentiable architecture search for GANs☆17Updated 3 years ago
- Zero-Shot Knowledge Distillation in Deep Networks in ICML2019☆49Updated 5 years ago
- Code for LIT, ICML 2019☆22Updated 5 years ago
- 3rd place solution for NeurIPS 2019 MicroNet challenge☆35Updated 4 years ago
- ICLR2019, Multilingual Neural Machine Translation with Knowledge Distillation☆70Updated 3 years ago
- Code for paper "Energy-Constrained Compression for Deep Neural Networks via Weighted Sparse Projection and Layer Input Masking"☆17Updated 5 years ago
- The implementation of multi-branch attentive Transformer (MAT).☆33Updated 4 years ago
- The collection of training tricks of binarized neural networks.☆71Updated 3 years ago