KohakuBlueleaf / guanaco-lora
Instruct-tune LLaMA on consumer hardware
☆72Updated last year
Alternatives and similar repositories for guanaco-lora:
Users that are interested in guanaco-lora are comparing it to the libraries listed below
- Train llama with lora on one 4090 and merge weight of lora to work as stanford alpaca.☆50Updated last year
- 📖 — Notebooks related to RWKV☆59Updated last year
- 4 bits quantization of LLaMa using GPTQ☆131Updated last year
- Inference script for Meta's LLaMA models using Hugging Face wrapper☆111Updated last year
- QLoRA: Efficient Finetuning of Quantized LLMs☆77Updated 10 months ago
- A light proxy solution for HuggingFace hub.☆46Updated last year
- Gradio UI for RWKV LLM☆28Updated last year
- MultilingualShareGPT, the free multi-language corpus for LLM training☆72Updated last year
- ☆52Updated last year
- ☆81Updated 9 months ago
- This project is established for real-time training of the RWKV model.☆49Updated 8 months ago
- zero零训练llm调参☆31Updated last year
- 8-bit CUDA functions for PyTorch☆44Updated last year
- 8-bit CUDA functions for PyTorch in Windows 10☆68Updated last year
- 4 bits quantization of SantaCoder using GPTQ☆51Updated last year
- Implementation of the Mamba SSM with hf_integration.☆56Updated 5 months ago
- This is a text generation method which returns a generator, streaming out each token in real-time during inference, based on Huggingface/…☆96Updated 11 months ago
- A high-throughput and memory-efficient inference and serving engine for LLMs☆130Updated 7 months ago
- Script and instruction how to fine-tune large RWKV model on your data for Alpaca dataset☆31Updated last year
- ☆33Updated 6 months ago
- RWKV infctx trainer, for training arbitary context sizes, to 10k and beyond!☆144Updated 6 months ago
- Tools for content datamining and NLP at scale☆42Updated 7 months ago
- Fast LLM Training CodeBase With dynamic strategy choosing [Deepspeed+Megatron+FlashAttention+CudaFusionKernel+Compiler];☆36Updated last year
- ChatGPT-like Web UI for RWKVstic☆100Updated last year
- Load multiple LoRA modules simultaneously and automatically switch the appropriate combination of LoRA modules to generate the best answe…☆149Updated last year
- 全球首个StableVicuna中文优化版。☆64Updated last year
- Modified Stanford-Alpaca Trainer for Training Replit's Code Model☆40Updated last year
- Enhancing LangChain prompts to work better with RWKV models☆34Updated last year
- a Fine-tuned LLaMA that is Good at Arithmetic Tasks☆177Updated last year
- RWKV is a RNN with transformer-level LLM performance. It can be directly trained like a GPT (parallelizable). So it's combining the best …☆410Updated last year