HazyResearch / flyLinks
☆208Updated 2 years ago
Alternatives and similar repositories for fly
Users that are interested in fly are comparing it to the libraries listed below
Sorting:
- Code for the NeurIPS 2022 paper "Optimal Brain Compression: A Framework for Accurate Post-Training Quantization and Pruning".☆122Updated last year
- ☆151Updated 2 years ago
- ☆147Updated 2 years ago
- A research library for pytorch-based neural network pruning, compression, and more.☆162Updated 2 years ago
- Block Sparse movement pruning☆80Updated 4 years ago
- ☆42Updated last year
- [NeurIPS 2022] A Fast Post-Training Pruning Framework for Transformers☆190Updated 2 years ago
- This repository contains integer operators on GPUs for PyTorch.☆205Updated last year
- This project is the official implementation of our accepted ICLR 2022 paper BiBERT: Accurate Fully Binarized BERT.☆88Updated 2 years ago
- Lightweight torch implementation of rigl, a sparse-to-sparse optimizer.☆57Updated 3 years ago
- Official Pytorch Implementation of "Outlier Weighed Layerwise Sparsity (OWL): A Missing Secret Sauce for Pruning LLMs to High Sparsity"☆69Updated 11 months ago
- Block-sparse primitives for PyTorch☆156Updated 4 years ago
- Soft Threshold Weight Reparameterization for Learnable Sparsity☆91Updated 2 years ago
- ☆20Updated last year
- ☆57Updated last year
- [NeurIPS'23] Speculative Decoding with Big Little Decoder☆92Updated last year
- [KDD'22] Learned Token Pruning for Transformers☆98Updated 2 years ago
- ☆157Updated last year
- ☆42Updated 2 years ago
- ☆204Updated 3 years ago
- Revisiting Efficient Training Algorithms For Transformer-based Language Models (NeurIPS 2023)☆80Updated last year
- Efficient LLM Inference Acceleration using Prompting☆48Updated 8 months ago
- Code repo for the paper BiT Robustly Binarized Multi-distilled Transformer☆108Updated last year
- [IJCAI'22 Survey] Recent Advances on Neural Network Pruning at Initialization.☆59Updated last year
- ☆236Updated 2 years ago
- [ICLR 2023] "Sparsity May Cry: Let Us Fail (Current) Sparse Neural Networks Together!" Shiwei Liu, Tianlong Chen, Zhenyu Zhang, Xuxi Chen…☆28Updated last year
- ☆10Updated 3 years ago
- This repository contains the experimental PyTorch native float8 training UX☆224Updated 10 months ago
- ☆105Updated 10 months ago
- The official PyTorch implementation of the NeurIPS2022 (spotlight) paper, Outlier Suppression: Pushing the Limit of Low-bit Transformer L…☆47Updated 2 years ago