johnsmith0031 / alpaca_lora_4bit
☆536Updated last year
Alternatives and similar repositories for alpaca_lora_4bit:
Users that are interested in alpaca_lora_4bit are comparing it to the libraries listed below
- ☆456Updated last year
- Tune any FALCON in 4-bit☆466Updated last year
- Landmark Attention: Random-Access Infinite Context Length for Transformers☆420Updated last year
- Customizable implementation of the self-instruct paper.☆1,034Updated 10 months ago
- ☆406Updated last year
- Landmark Attention: Random-Access Infinite Context Length for Transformers QLoRA☆123Updated last year
- Finetuning Large Language Models on One Consumer GPU in 2 Bits☆714Updated 8 months ago
- Alpaca dataset from Stanford, cleaned and curated☆1,532Updated last year
- LaMini-LM: A Diverse Herd of Distilled Models from Large-Scale Instructions☆815Updated last year
- Falcon LLM ggml framework with CPU and GPU support☆246Updated last year
- RWKV is a RNN with transformer-level LLM performance. It can be directly trained like a GPT (parallelizable). So it's combining the best …☆410Updated last year
- Official repository for LongChat and LongEval☆519Updated 8 months ago
- ☆538Updated last month
- 4 bits quantization of LLaMA using GPTQ☆3,032Updated 6 months ago
- Code for fine-tuning Platypus fam LLMs using LoRA☆626Updated 11 months ago
- This repository contains code for extending the Stanford Alpaca synthetic instruction tuning to existing instruction-tuned models such as…☆351Updated last year
- C++ implementation for BLOOM☆810Updated last year
- SoTA Transformers with C-backend for fast inference on your CPU.☆312Updated last year
- C++ implementation for 💫StarCoder☆450Updated last year
- [ICML 2024] SqueezeLLM: Dense-and-Sparse Quantization☆669Updated 5 months ago
- 4 bits quantization of LLaMa using GPTQ☆131Updated last year
- Repo for fine-tuning Casual LLMs☆454Updated 10 months ago
- Fast Inference Solutions for BLOOM☆563Updated 3 months ago
- LLaMa retrieval plugin script using OpenAI's retrieval plugin☆324Updated last year
- A torchless, c++ rwkv implementation using 8bit quantization, written in cuda/hip/vulkan for maximum compatibility and minimum dependenci…☆309Updated 11 months ago
- Merge Transformers language models by use of gradient parameters.☆203Updated 5 months ago
- fastLLaMa: An experimental high-performance framework for running Decoder-only LLMs with 4-bit quantization in Python using a C/C++ backe…☆408Updated last year
- ggml implementation of BERT☆476Updated 11 months ago
- batched loras☆338Updated last year
- Official code for ReLoRA from the paper Stack More Layers Differently: High-Rank Training Through Low-Rank Updates☆440Updated 9 months ago