BorealisAI / neuzip
Official repository for the paper "NeuZip: Memory-Efficient Training and Inference with Dynamic Compression of Neural Networks". This repository contains the code for the experiments in the paper.
☆32Updated 3 weeks ago
Related projects ⓘ
Alternatives and complementary repositories for neuzip
- Official implementation of ECCV24 paper: POA☆24Updated 3 months ago
- Official repository for the paper "SwitchHead: Accelerating Transformers with Mixture-of-Experts Attention"☆92Updated last month
- Lottery Ticket Adaptation☆36Updated last month
- FBI-LLM: Scaling Up Fully Binarized LLMs from Scratch via Autoregressive Distillation☆46Updated 4 months ago
- ☆35Updated 9 months ago
- A repository for research on medium sized language models.☆74Updated 5 months ago
- ☆43Updated 4 months ago
- This repository contains code for the MicroAdam paper.☆12Updated 4 months ago
- The code repository for the CURLoRA research paper. Stable LLM continual fine-tuning and catastrophic forgetting mitigation.☆38Updated 2 months ago
- Zeta implementation of a reusable and plug in and play feedforward from the paper "Exponentially Faster Language Modeling"☆15Updated last week
- ☆63Updated last month
- A list of language models with permissive licenses such as MIT or Apache 2.0☆24Updated 2 weeks ago
- ☆57Updated last week
- QuIP quantization☆46Updated 8 months ago
- ☆59Updated last month
- This repository includes the code to download the curated HuggingFace papers into a single markdown formatted file☆14Updated 3 months ago
- Official implementation of the paper "MMInA: Benchmarking Multihop Multimodal Internet Agents"☆38Updated 7 months ago
- A public implementation of the ReLoRA pretraining method, built on Lightning-AI's Pytorch Lightning suite.☆33Updated 8 months ago
- Here we will test various linear attention designs.☆56Updated 6 months ago
- ☆18Updated last month
- Recaption large (Web)Datasets with vllm and save the artifacts.☆30Updated last month
- ☆21Updated this week
- Repository for CPU Kernel Generation for LLM Inference☆25Updated last year
- My Implementation of Q-Sparse: All Large Language Models can be Fully Sparsely-Activated☆30Updated 3 months ago
- Official repository for ICML 2024 paper "MoRe Fine-Tuning with 10x Fewer Parameters"☆16Updated 2 weeks ago
- From GaLore to WeLore: How Low-Rank Weights Non-uniformly Emerge from Low-Rank Gradients. Ajay Jaiswal, Lu Yin, Zhenyu Zhang, Shiwei Liu,…☆43Updated 4 months ago
- Repository for Sparse Finetuning of LLMs via modified version of the MosaicML llmfoundry☆38Updated 10 months ago
- A toolkit for fine-tuning, inferencing, and evaluating GreenBitAI's LLMs.☆74Updated last month
- Enhancement in Multimodal Representation Learning.☆39Updated 8 months ago
- Implementation of the paper: "Mixture-of-Depths: Dynamically allocating compute in transformer-based language models"☆73Updated this week