KohakuBlueleaf / guanaco-lora
Instruct-tune LLaMA on consumer hardware
☆73Updated last year
Related projects ⓘ
Alternatives and complementary repositories for guanaco-lora
- An unsupervised model merging algorithm for Transformers-based language models.☆99Updated 6 months ago
- 4 bits quantization of LLaMa using GPTQ☆130Updated last year
- QLoRA: Efficient Finetuning of Quantized LLMs☆77Updated 6 months ago
- Train llama with lora on one 4090 and merge weight of lora to work as stanford alpaca.☆50Updated last year
- Demonstration that finetuning RoPE model on larger sequences than the pre-trained model adapts the model context limit☆63Updated last year
- 4 bits quantization of SantaCoder using GPTQ☆53Updated last year
- 8-bit CUDA functions for PyTorch in Windows 10☆71Updated last year
- 8-bit CUDA functions for PyTorch☆45Updated last year
- ChatGPT-like Web UI for RWKVstic☆100Updated last year
- Image Diffusion block merging technique applied to transformers based Language Models.☆54Updated last year
- Conversational Language model toolkit for training against human preferences.☆40Updated 7 months ago
- Script and instruction how to fine-tune large RWKV model on your data for Alpaca dataset☆31Updated last year
- ☆33Updated 3 months ago
- ☆42Updated last year
- ☆27Updated last year
- This project is established for real-time training of the RWKV model.☆50Updated 5 months ago
- Inference script for Meta's LLaMA models using Hugging Face wrapper☆111Updated last year
- Train Llama Loras Easily☆29Updated last year
- 📖 — Notebooks related to RWKV☆59Updated last year
- ☆79Updated 11 months ago
- A high-throughput and memory-efficient inference and serving engine for LLMs☆129Updated 4 months ago
- Tune MPTs☆84Updated last year
- GPTQLoRA: Efficient Finetuning of Quantized LLMs with GPTQ☆96Updated last year
- RWKV infctx trainer, for training arbitary context sizes, to 10k and beyond!☆133Updated 2 months ago
- Landmark Attention: Random-Access Infinite Context Length for Transformers QLoRA☆124Updated last year
- A pipeline parallel training script for LLMs.☆83Updated 3 weeks ago
- Finetune any model on HF in less than 30 seconds☆56Updated this week
- Framework agnostic python runtime for RWKV models☆145Updated last year
- Experimental sampler to make LLMs more creative☆30Updated last year