lxe / llama-peft-tuner
Tune LLaMa-7B on Alpaca Dataset using PEFT / LORA Based on @zphang's https://github.com/zphang/minimal-llama scripts.
☆24Updated last year
Alternatives and similar repositories for llama-peft-tuner:
Users that are interested in llama-peft-tuner are comparing it to the libraries listed below
- Instruct-tune Open LLaMA / RedPajama / StableLM models on consumer hardware using QLoRA☆80Updated last year
- QLoRA: Efficient Finetuning of Quantized LLMs☆77Updated 10 months ago
- ☆96Updated last year
- Code repository for the c-BTM paper☆105Updated last year
- Finetune Falcon, LLaMA, MPT, and RedPajama on consumer hardware using PEFT LoRA☆101Updated 6 months ago
- Exploring finetuning public checkpoints on filter 8K sequences on Pile☆115Updated last year
- Reimplementation of the task generation part from the Alpaca paper☆119Updated last year
- Spherical Merge Pytorch/HF format Language Models with minimal feature loss.☆115Updated last year
- Experiments with generating opensource language model assistants☆97Updated last year
- ☆94Updated last year
- Minimal code to train a Large Language Model (LLM).☆168Updated 2 years ago
- Reverse Instructions to generate instruction tuning data with corpus examples☆208Updated 11 months ago
- [AAAI 2024] Investigating the Effectiveness of Task-Agnostic Prefix Prompt for Instruction Following☆79Updated 5 months ago
- Multipack distributed sampler for fast padding-free training of LLMs☆184Updated 6 months ago
- Code and model release for the paper "Task-aware Retrieval with Instructions" by Asai et al.☆162Updated last year
- Patch for MPT-7B which allows using and training a LoRA☆58Updated last year
- This is the repo for the paper Shepherd -- A Critic for Language Model Generation☆218Updated last year
- Open Instruction Generalist is an assistant trained on massive synthetic instructions to perform many millions of tasks☆208Updated last year
- Official implementation for 'Extending LLMs’ Context Window with 100 Samples'☆76Updated last year
- Here is a Google Colab Notebook for fine-tuning Alpaca Lora (within 3 hours with a 40GB A100 GPU)☆38Updated last year
- Inference script for Meta's LLaMA models using Hugging Face wrapper☆111Updated last year
- ☆177Updated last year
- Fine-tuning 6-Billion GPT-J (& other models) with LoRA and 8-bit compression☆66Updated 2 years ago
- Implementation of the paper: "Answering Questions by Meta-Reasoning over Multiple Chains of Thought"☆93Updated last year
- [ICLR 2023] Guess the Instruction! Flipped Learning Makes Language Models Stronger Zero-Shot Learners☆113Updated 5 months ago
- ☆69Updated last year
- Tk-Instruct is a Transformer model that is tuned to solve many NLP tasks by following instructions.☆180Updated 2 years ago
- ☆160Updated last year
- Official code for ACL 2023 (short, findings) paper "Recursion of Thought: A Divide and Conquer Approach to Multi-Context Reasoning with L…☆42Updated last year
- This repository contains code for extending the Stanford Alpaca synthetic instruction tuning to existing instruction-tuned models such as…☆350Updated last year