skolai / fewbitLinks
Compression schema for gradients of activations in backward pass
☆44Updated 2 years ago
Alternatives and similar repositories for fewbit
Users that are interested in fewbit are comparing it to the libraries listed below
Sorting:
- Learning to Initialize Neural Networks for Stable and Efficient Training☆139Updated 3 years ago
- PyTorch implementation of L2L execution algorithm☆108Updated 2 years ago
- Experiment of using Tangent to autodiff triton☆80Updated last year
- ☆213Updated 2 years ago
- A library for unit scaling in PyTorch☆129Updated last month
- ☆37Updated 2 weeks ago
- Latest Weight Averaging (NeurIPS HITY 2022)☆31Updated 2 years ago
- A block oriented training approach for inference time optimization.☆34Updated last year
- ☆29Updated 2 years ago
- ☆75Updated 2 years ago
- ☆118Updated last year
- Customized matrix multiplication kernels☆56Updated 3 years ago
- Official implementation of the paper "You Do Not Fully Utilize Transformer's Representation Capacity"☆30Updated 3 months ago
- PyTorch implementation of HashedNets☆36Updated 2 years ago
- ☆60Updated 5 years ago
- Little article showing how to load pytorch's models with linear memory consumption☆34Updated 3 years ago
- ☆36Updated 8 months ago
- The official implementation of the ChordMixer architecture.☆61Updated 2 years ago
- Demo of the unit_scaling library, showing how a model can be easily adapted to train in FP8.☆46Updated last year
- ☆18Updated 4 months ago
- ☆51Updated last year
- Hacks for PyTorch☆19Updated 2 years ago
- Revisiting Efficient Training Algorithms For Transformer-based Language Models (NeurIPS 2023)☆81Updated 2 years ago
- Dynamic Neural Architecture Search Toolkit☆30Updated 8 months ago
- Implementation of "Gradients without backpropagation" paper (https://arxiv.org/abs/2202.08587) using functorch☆111Updated 2 years ago
- Layerwise Batch Entropy Regularization☆23Updated 3 years ago
- Implementation of fused cosine similarity attention in the same style as Flash Attention☆215Updated 2 years ago
- Implementation of a Transformer using ReLA (Rectified Linear Attention) from https://arxiv.org/abs/2104.07012☆49Updated 3 years ago
- ☆159Updated last year
- Official code for "Distributed Deep Learning in Open Collaborations" (NeurIPS 2021)☆117Updated 3 years ago