GATECH-EIC / FracTrain
[NeurIPS 2020] "FracTrain: Fractionally Squeezing Bit Savings Both Temporally and Spatially for Efficient DNN Training" by Yonggan Fu, Haoran You, Yang Zhao, Yue Wang, Chaojian Li, Kailash Gopalakrishnan, Zhangyang Wang, Yingyan Lin
☆11Updated 3 years ago
Alternatives and similar repositories for FracTrain:
Users that are interested in FracTrain are comparing it to the libraries listed below
- Official implementation of "Searching for Winograd-aware Quantized Networks" (MLSys'20)☆27Updated last year
- BSQ: Exploring Bit-Level Sparsity for Mixed-Precision Neural Network Quantization (ICLR 2021)☆40Updated 4 years ago
- ☆18Updated 3 years ago
- Code for ICML 2021 submission☆35Updated 3 years ago
- [ICML 2021] "Auto-NBA: Efficient and Effective Search Over the Joint Space of Networks, Bitwidths, and Accelerators" by Yonggan Fu, Yonga…☆15Updated 3 years ago
- [ICML 2021] "Double-Win Quant: Aggressively Winning Robustness of Quantized DeepNeural Networks via Random Precision Training and Inferen…☆13Updated 3 years ago
- Simulator for BitFusion☆97Updated 4 years ago
- Neural Network Quantization With Fractional Bit-widths☆12Updated 4 years ago
- BitPack is a practical tool to efficiently save ultra-low precision/mixed-precision quantized models.☆51Updated 2 years ago
- PyTorch implementation of EdMIPS: https://arxiv.org/pdf/2004.05795.pdf☆58Updated 4 years ago
- MNSIM_Python_v1.0. The former circuits-level version link: https://github.com/Zhu-Zhenhua/MNSIM_V1.1☆34Updated last year
- ☆33Updated 3 years ago
- A collection of research papers on efficient training of DNNs☆70Updated 2 years ago
- TBNv2: Convolutional Neural Network With Ternary Inputs and Binary Weights☆17Updated 5 years ago
- Measuring and predicting on-device metrics (latency, power, etc.) of machine learning models☆66Updated last year
- pytorch fixed point training tool/framework☆34Updated 4 years ago
- ☆32Updated 4 years ago
- Post-training sparsity-aware quantization☆34Updated 2 years ago
- [NeurIPS 2023] ShiftAddViT: Mixture of Multiplication Primitives Towards Efficient Vision Transformer☆32Updated last year
- ☆23Updated 3 months ago
- A Out-of-box PyTorch Scaffold for Neural Network Quantization-Aware-Training (QAT) Research. Website: https://github.com/zhutmost/neuralz…☆26Updated 2 years ago
- ☆39Updated 2 years ago
- AFP is a hardware-friendly quantization framework for DNNs, which is contributed by Fangxin Liu and Wenbo Zhao.☆12Updated 3 years ago
- Official implementation of EMNLP'23 paper "Revisiting Block-based Quantisation: What is Important for Sub-8-bit LLM Inference?"☆19Updated last year
- ☆19Updated 4 years ago
- ☆43Updated last year
- Here are some implementations of basic hardware units in RTL language (verilog for now), which can be used for area/power evaluation and …☆11Updated last year
- ☆92Updated last year