AlpinDale / QuIP-for-Llama
Code for paper: "QuIP: 2-Bit Quantization of Large Language Models With Guarantees" adapted for Llama models
☆35Updated last year
Alternatives and similar repositories for QuIP-for-Llama:
Users that are interested in QuIP-for-Llama are comparing it to the libraries listed below
- QuIP quantization☆52Updated last year
- GPTQLoRA: Efficient Finetuning of Quantized LLMs with GPTQ☆103Updated last year
- ☆131Updated last month
- Advanced Ultra-Low Bitrate Compression Techniques for the LLaMA Family of LLMs☆110Updated last year
- A toolkit for fine-tuning, inferencing, and evaluating GreenBitAI's LLMs.☆83Updated last month
- ☆73Updated last year
- PB-LLM: Partially Binarized Large Language Models☆152Updated last year
- An efficent implementation of the method proposed in "The Era of 1-bit LLMs"☆154Updated 6 months ago
- RWKV-7: Surpassing GPT☆83Updated 5 months ago
- Repository for Sparse Finetuning of LLMs via modified version of the MosaicML llmfoundry☆40Updated last year
- ☆118Updated last year
- ☆50Updated 6 months ago
- Code for paper: "QuIP: 2-Bit Quantization of Large Language Models With Guarantees"