KohakuBlueleaf / guanaco-lora
Instruct-tune LLaMA on consumer hardware
☆74Updated last year
Alternatives and similar repositories for guanaco-lora:
Users that are interested in guanaco-lora are comparing it to the libraries listed below
- 8-bit CUDA functions for PyTorch☆44Updated last year
- 4 bits quantization of LLaMa using GPTQ☆130Updated last year
- ☆82Updated 11 months ago
- 4 bits quantization of SantaCoder using GPTQ☆51Updated last year
- QLoRA: Efficient Finetuning of Quantized LLMs☆78Updated last year
- ChatGPT-like Web UI for RWKVstic☆100Updated 2 years ago
- ☆42Updated 2 years ago
- Image Diffusion block merging technique applied to transformers based Language Models.☆54Updated last year
- A finetuning pipeline for instruct tuning Raven 14bn using QLORA 4bit and the Ditty finetuning library☆28Updated 10 months ago
- ☆54Updated 2 years ago
- 📖 — Notebooks related to RWKV☆59Updated last year
- Train llama with lora on one 4090 and merge weight of lora to work as stanford alpaca.☆51Updated last year
- This project is established for real-time training of the RWKV model.☆49Updated 11 months ago
- Framework agnostic python runtime for RWKV models☆146Updated last year
- BigKnow2022: Bringing Language Models Up to Speed☆15Updated 2 years ago
- ☆26Updated 2 years ago
- Demonstration that finetuning RoPE model on larger sequences than the pre-trained model adapts the model context limit☆63Updated last year
- ☆12Updated last year
- 8-bit CUDA functions for PyTorch in Windows 10☆69Updated last year
- Instruct-tuning LLaMA on consumer hardware☆66Updated 2 years ago
- SparseGPT + GPTQ Compression of LLMs like LLaMa, OPT, Pythia☆41Updated 2 years ago
- An unsupervised model merging algorithm for Transformers-based language models.☆105Updated 11 months ago
- ☆34Updated 9 months ago
- Finetune any model on HF in less than 30 seconds☆58Updated 2 weeks ago
- Zeus LLM Trainer is a rewrite of Stanford Alpaca aiming to be the trainer for all Large Language Models☆69Updated last year
- Script and instruction how to fine-tune large RWKV model on your data for Alpaca dataset☆31Updated 2 years ago
- BlinkDL's RWKV-v4 running in the browser☆47Updated 2 years ago
- Landmark Attention: Random-Access Infinite Context Length for Transformers QLoRA☆123Updated last year
- This project aims to make RWKV Accessible to everyone using a Hugging Face like interface, while keeping it close to the R and D RWKV bra…☆64Updated last year
- Inference script for Meta's LLaMA models using Hugging Face wrapper☆110Updated 2 years ago