Low-bit optimizers for PyTorch
☆138Oct 9, 2023Updated 2 years ago
Alternatives and similar repositories for low-bit-optimizers
Users that are interested in low-bit-optimizers are comparing it to the libraries listed below
Sorting:
- A Tight-fisted Optimizer☆50Mar 7, 2023Updated 2 years ago
- Pytorch implementation of "Oscillation-Reduced MXFP4 Training for Vision Transformers" on DeiT Model Pre-training☆36Jun 20, 2025Updated 8 months ago
- code for Scaling Laws of RoPE-based Extrapolation☆73Oct 16, 2023Updated 2 years ago
- See https://github.com/cuda-mode/triton-index/ instead!☆11May 8, 2024Updated last year
- Code for the paper "QMoE: Practical Sub-1-Bit Compression of Trillion-Parameter Models".☆279Nov 3, 2023Updated 2 years ago
- Demonstration that finetuning RoPE model on larger sequences than the pre-trained model adapts the model context limit☆63Jun 21, 2023Updated 2 years ago
- ☆156Jun 22, 2023Updated 2 years ago
- ☆14Aug 1, 2025Updated 7 months ago
- HALO: Hadamard-Assisted Low-Precision Optimization and Training method for finetuning LLMs. 🚀 The official implementation of https://arx…☆29Feb 17, 2025Updated last year
- ☆63Jul 21, 2024Updated last year
- Code for Blog Post: Can Better Cold-Start Strategies Improve RL Training for LLMs?☆19Mar 9, 2025Updated 11 months ago
- ☆235Jun 11, 2024Updated last year
- GaLore: Memory-Efficient LLM Training by Gradient Low-Rank Projection☆1,677Oct 28, 2024Updated last year
- Linear Attention Sequence Parallelism (LASP)☆88Jun 4, 2024Updated last year
- Revisiting Efficient Training Algorithms For Transformer-based Language Models (NeurIPS 2023)☆81Aug 30, 2023Updated 2 years ago
- Microsoft Automatic Mixed Precision Library☆636Dec 1, 2025Updated 3 months ago
- LOMO: LOw-Memory Optimization☆988Jul 2, 2024Updated last year
- GPTQLoRA: Efficient Finetuning of Quantized LLMs with GPTQ☆101May 30, 2023Updated 2 years ago
- [Neurips 2022] “ Back Razor: Memory-Efficient Transfer Learning by Self-Sparsified Backpropogation”, Ziyu Jiang*, Xuxi Chen*, Xueqin Huan…☆19Mar 14, 2023Updated 2 years ago
- Official implementation of ICML 2024 paper "ExCP: Extreme LLM Checkpoint Compression via Weight-Momentum Joint Shrinking".☆47Jul 12, 2024Updated last year
- ☆22Jul 27, 2023Updated 2 years ago
- ☆20Nov 3, 2024Updated last year
- LLMs as Collaboratively Edited Knowledge Bases☆46Feb 8, 2026Updated 3 weeks ago
- The official implementation of the EMNLP 2023 paper LLM-FP4☆220Dec 15, 2023Updated 2 years ago
- Continual Resilient (CoRe) Optimizer for PyTorch☆11Jun 10, 2024Updated last year
- [ICLR 2025] COAT: Compressing Optimizer States and Activation for Memory-Efficient FP8 Training☆260Aug 9, 2025Updated 6 months ago
- Fast and Robust Early-Exiting Framework for Autoregressive Language Models with Synchronized Parallel Decoding (EMNLP 2023 Long)☆65Sep 28, 2024Updated last year
- Ring attention implementation with flash attention☆986Sep 10, 2025Updated 5 months ago
- Reorder-based post-training quantization for large language model☆199May 17, 2023Updated 2 years ago
- Collaborative Training of Large Language Models in an Efficient Way☆419Aug 28, 2024Updated last year
- AudioSR-Upsampling (any -> 48kHz)☆42Feb 13, 2024Updated 2 years ago
- Unofficial implementation of AlpaGasus☆94Sep 23, 2023Updated 2 years ago
- ☆120Jan 8, 2026Updated last month
- ☆21Mar 3, 2025Updated 11 months ago
- Official code for our paper, "LoRA-Pro: Are Low-Rank Adapters Properly Optimized? "☆144Apr 8, 2025Updated 10 months ago
- Code for Neurips24 paper: QuaRot, an end-to-end 4-bit inference of large language models.☆485Nov 26, 2024Updated last year
- Contrastive Chain-of-Thought Prompting☆68Nov 18, 2023Updated 2 years ago
- [ICML 2024] SqueezeLLM: Dense-and-Sparse Quantization☆712Aug 13, 2024Updated last year
- ☆553Feb 8, 2026Updated 3 weeks ago