BorealisAI / flora-opt
This is the official repository for the paper "Flora: Low-Rank Adapters Are Secretly Gradient Compressors" in ICML 2024.
☆97Updated 7 months ago
Alternatives and similar repositories for flora-opt:
Users that are interested in flora-opt are comparing it to the libraries listed below
- ☆66Updated 7 months ago
- ☆125Updated last year
- ☆166Updated last year
- ☆79Updated 3 months ago
- ☆71Updated 6 months ago
- ☆192Updated 2 months ago
- Implementation of 🥥 Coconut, Chain of Continuous Thought, in Pytorch☆156Updated last month
- Token Omission Via Attention☆123Updated 4 months ago
- Pytorch implementation of the PEER block from the paper, Mixture of A Million Experts, by Xu Owen He at Deepmind☆118Updated 5 months ago
- ☆75Updated last month
- Code accompanying the paper "Massive Activations in Large Language Models"☆140Updated 11 months ago
- Function Vectors in Large Language Models (ICLR 2024)☆138Updated 4 months ago
- [ICLR 2024 Spotlight] Code for the paper "Merge, Then Compress: Demystify Efficient SMoE with Hints from Its Routing Policy"☆72Updated 8 months ago
- Activation-aware Singular Value Decomposition for Compressing Large Language Models☆56Updated 3 months ago
- A brief and partial summary of RLHF algorithms.☆93Updated 2 months ago
- The simplest, fastest repository for training/finetuning medium-sized GPTs.☆95Updated 3 months ago
- Model Stock: All we need is just a few fine-tuned models☆102Updated 4 months ago
- Code to reproduce "Transformers Can Do Arithmetic with the Right Embeddings", McLeish et al (NeurIPS 2024)☆185Updated 8 months ago
- [NeurIPS 2024] Goldfish Loss: Mitigating Memorization in Generative LLMs☆82Updated 3 months ago
- [NeurIPS 24 Spotlight] MaskLLM: Learnable Semi-structured Sparsity for Large Language Models