huggingface / nn_pruningLinks
Prune a model while finetuning or training.
☆403Updated 3 years ago
Alternatives and similar repositories for nn_pruning
Users that are interested in nn_pruning are comparing it to the libraries listed below
Sorting:
- Library for 8-bit optimizers and quantization routines.☆715Updated 2 years ago
- FastFormers - highly efficient transformer models for NLU☆705Updated 3 months ago
- Parallelformers: An Efficient Model Parallelization Toolkit for Deployment☆790Updated 2 years ago
- Repository containing code for "How to Train BERT with an Academic Budget" paper☆313Updated last year
- A library for researching neural networks compression and acceleration methods.☆138Updated 9 months ago
- [NeurIPS 2022] A Fast Post-Training Pruning Framework for Transformers☆190Updated 2 years ago
- [ACL 2022] Structured Pruning Learns Compact and Accurate Models https://arxiv.org/abs/2204.00408☆195Updated 2 years ago
- Implementation of a Transformer, but completely in Triton☆268Updated 3 years ago
- [ICML'21 Oral] I-BERT: Integer-only BERT Quantization☆250Updated 2 years ago
- Fast Block Sparse Matrices for Pytorch☆547Updated 4 years ago
- [ACL'20] HAT: Hardware-Aware Transformers for Efficient Natural Language Processing☆335Updated 11 months ago
- ☆204Updated 3 years ago
- Understanding the Difficulty of Training Transformers☆329Updated 3 years ago
- Running BERT without Padding☆471Updated 3 years ago
- An efficient implementation of the popular sequence models for text generation, summarization, and translation tasks. https://arxiv.org/p…☆433Updated 2 years ago
- Flexible components pairing 🤗 Transformers with Pytorch Lightning☆609Updated 2 years ago
- Root Mean Square Layer Normalization☆242Updated 2 years ago
- GPTQ inference Triton kernel☆302Updated 2 years ago
- Accelerate PyTorch models with ONNX Runtime☆362Updated 4 months ago
- Code repo for the paper "LLM-QAT Data-Free Quantization Aware Training for Large Language Models"☆300Updated 3 months ago
- A GPU performance profiling tool for PyTorch models☆503Updated 3 years ago
- ☆208Updated 2 years ago
- DiffQ performs differentiable quantization using pseudo quantization noise. It can automatically tune the number of bits used per weight …☆235Updated 2 years ago
- Sequence modeling with Mega.☆296Updated 2 years ago
- Official PyTorch Implementation of Long-Short Transformer (NeurIPS 2021).☆225Updated 3 years ago
- This is a repository with the code for the ACL 2019 paper "Analyzing Multi-Head Self-Attention: Specialized Heads Do the Heavy Lifting, t…☆314Updated 3 years ago
- ☆250Updated 11 months ago
- DeeBERT: Dynamic Early Exiting for Accelerating BERT Inference☆156Updated 3 years ago
- ☆411Updated last year
- Fully featured implementation of Routing Transformer☆295Updated 3 years ago