leehanchung / lora-instructLinks
Finetune Falcon, LLaMA, MPT, and RedPajama on consumer hardware using PEFT LoRA
☆104Updated 8 months ago
Alternatives and similar repositories for lora-instruct
Users that are interested in lora-instruct are comparing it to the libraries listed below
Sorting:
- QLoRA: Efficient Finetuning of Quantized LLMs☆79Updated last year
- Notus is a collection of fine-tuned LLMs using SFT, DPO, SFT+DPO, and/or any other RLHF techniques, while always keeping a data-first app…☆169Updated 2 years ago
- Small and Efficient Mathematical Reasoning LLMs☆73Updated 2 years ago
- Reimplementation of the task generation part from the Alpaca paper☆119Updated 2 years ago
- ☆86Updated 2 years ago
- Multi-Domain Expert Learning☆67Updated 2 years ago
- ☆23Updated 2 years ago
- Mixing Language Models with Self-Verification and Meta-Verification☆112Updated last year
- ☆78Updated 2 years ago
- Exploring finetuning public checkpoints on filter 8K sequences on Pile☆116Updated 2 years ago
- QLoRA with Enhanced Multi GPU Support☆37Updated 2 years ago
- Zeus LLM Trainer is a rewrite of Stanford Alpaca aiming to be the trainer for all Large Language Models☆70Updated 2 years ago
- Tune MPTs☆84Updated 2 years ago
- Code repository for the c-BTM paper☆108Updated 2 years ago
- GPTQLoRA: Efficient Finetuning of Quantized LLMs with GPTQ☆101Updated 2 years ago
- ModuleFormer is a MoE-based architecture that includes two different types of experts: stick-breaking attention heads and feedforward exp…☆226Updated 4 months ago
- ☆37Updated 2 years ago
- Some simple scripts that I use day-to-day when working with LLMs and Huggingface Hub☆161Updated 2 years ago
- Weekly visualization report of Open LLM model performance based on 4 metrics.☆86Updated 2 years ago
- Spherical Merge Pytorch/HF format Language Models with minimal feature loss.☆143Updated 2 years ago
- Evaluating LLMs with CommonGen-Lite☆93Updated last year
- Just a bunch of benchmark logs for different LLMs☆119Updated last year
- ☆95Updated 2 years ago
- The GeoV model is a large langauge model designed by Georges Harik and uses Rotary Positional Embeddings with Relative distances (RoPER).…☆121Updated 2 years ago
- TART: A plug-and-play Transformer module for task-agnostic reasoning☆202Updated 2 years ago
- Convenient wrapper for fine-tuning and inference of Large Language Models (LLMs) with several quantization techniques (GTPQ, bitsandbytes…☆146Updated 2 years ago
- Sakura-SOLAR-DPO: Merge, SFT, and DPO☆116Updated 2 years ago
- manage histories of LLM applied applications☆91Updated 2 years ago
- Official repo for the paper PHUDGE: Phi-3 as Scalable Judge. Evaluate your LLMs with or without custom rubric, reference answer, absolute…☆51Updated last year
- ☆74Updated 2 years ago