georgesung / llm_qlora
Fine-tuning LLMs using QLoRA
☆251Updated 10 months ago
Alternatives and similar repositories for llm_qlora:
Users that are interested in llm_qlora are comparing it to the libraries listed below
- Tune any FALCON in 4-bit☆466Updated last year
- A bagel, with everything.☆320Updated last year
- Fine-tune mistral-7B on 3090s, a100s, h100s☆710Updated last year
- Some simple scripts that I use day-to-day when working with LLMs and Huggingface Hub☆160Updated last year
- Customizable implementation of the self-instruct paper.☆1,043Updated last year
- This is our own implementation of 'Layer Selective Rank Reduction'☆235Updated 11 months ago
- Landmark Attention: Random-Access Infinite Context Length for Transformers☆422Updated last year
- Fully fine-tune large models like Mistral, Llama-2-13B, or Qwen-14B completely for free☆231Updated 5 months ago
- Convenient wrapper for fine-tuning and inference of Large Language Models (LLMs) with several quantization techniques (GTPQ, bitsandbytes…☆147Updated last year
- QLoRA: Efficient Finetuning of Quantized LLMs☆78Updated last year
- ☆168Updated last year
- The code we currently use to fine-tune models.☆114Updated 11 months ago
- This repository contains code for extending the Stanford Alpaca synthetic instruction tuning to existing instruction-tuned models such as…☆351Updated last year
- ☆122Updated last year
- Merge Transformers language models by use of gradient parameters.☆206Updated 8 months ago
- ☆153Updated 9 months ago
- Extend existing LLMs way beyond the original training length with constant memory usage, without retraining☆692Updated last year
- ☆412Updated last year
- ☆535Updated last year
- Notus is a collection of fine-tuned LLMs using SFT, DPO, SFT+DPO, and/or any other RLHF techniques, while always keeping a data-first app…☆167Updated last year
- Inference code for Mistral and Mixtral hacked up into original Llama implementation☆371Updated last year
- Falcon LLM ggml framework with CPU and GPU support☆246Updated last year
- A joint community effort to create one central leaderboard for LLMs.☆295Updated 8 months ago
- batched loras☆341Updated last year
- Generate textbook-quality synthetic LLM pretraining data☆498Updated last year
- Guide for fine-tuning Llama/Mistral/CodeLlama models and more☆583Updated this week
- TheBloke's Dockerfiles☆303Updated last year
- Code for fine-tuning Platypus fam LLMs using LoRA☆629Updated last year
- Small finetuned LLMs for a diverse set of useful tasks☆126Updated last year
- ☆94Updated last year