lxe / llama-peft-tuner
Tune LLaMa-7B on Alpaca Dataset using PEFT / LORA Based on @zphang's https://github.com/zphang/minimal-llama scripts.
☆25Updated 2 years ago
Alternatives and similar repositories for llama-peft-tuner
Users that are interested in llama-peft-tuner are comparing it to the libraries listed below
Sorting:
- Instruct-tune Open LLaMA / RedPajama / StableLM models on consumer hardware using QLoRA☆81Updated last year
- [AAAI 2024] Investigating the Effectiveness of Task-Agnostic Prefix Prompt for Instruction Following☆79Updated 8 months ago
- LLaMa Tuning with Stanford Alpaca Dataset using Deepspeed and Transformers☆51Updated 2 years ago
- ☆97Updated last year
- An experimental implementation of the retrieval-enhanced language model☆74Updated 2 years ago
- QLoRA: Efficient Finetuning of Quantized LLMs☆78Updated last year
- Official codebase for "SelFee: Iterative Self-Revising LLM Empowered by Self-Feedback Generation"☆226Updated last year
- Spherical Merge Pytorch/HF format Language Models with minimal feature loss.☆121Updated last year
- ☆179Updated 2 years ago
- Official code for ACL 2023 (short, findings) paper "Recursion of Thought: A Divide and Conquer Approach to Multi-Context Reasoning with L…☆43Updated last year
- Exploring finetuning public checkpoints on filter 8K sequences on Pile☆115Updated 2 years ago
- Finetune Falcon, LLaMA, MPT, and RedPajama on consumer hardware using PEFT LoRA☆103Updated 9 months ago
- Reverse Instructions to generate instruction tuning data with corpus examples☆211Updated last year
- Alpaca-lora for huggingface implementation using Deepspeed and FullyShardedDataParallel☆24Updated 2 years ago
- ☆51Updated last year
- Prompt tuning toolkit for GPT-2 and GPT-Neo☆88Updated 3 years ago
- Multipack distributed sampler for fast padding-free training of LLMs☆189Updated 9 months ago
- ☆67Updated last year
- Patch for MPT-7B which allows using and training a LoRA☆58Updated last year
- Instruct-tuning LLaMA on consumer hardware☆66Updated 2 years ago
- Implementation of Reinforcement Learning from Human Feedback (RLHF)☆173Updated 2 years ago
- Small and Efficient Mathematical Reasoning LLMs☆71Updated last year
- Here is a Google Colab Notebook for fine-tuning Alpaca Lora (within 3 hours with a 40GB A100 GPU)☆38Updated 2 years ago
- All available datasets for Instruction Tuning of Large Language Models☆250Updated last year
- About The corresponding code from our paper " REFINER: Reasoning Feedback on Intermediate Representations" (EACL 2024). Do not hesitate t…☆70Updated last year
- ☆106Updated last year
- This is the repo for the paper Shepherd -- A Critic for Language Model Generation☆219Updated last year
- 🥤🧑🏻🚀Code and dataset for our EMNLP 2023 paper - "SODA: Million-scale Dialogue Distillation with Social Commonsense Contextualization…☆230Updated last year
- Fine-tuning 6-Billion GPT-J (& other models) with LoRA and 8-bit compression☆66Updated 2 years ago
- Inference code for facebook LLaMA models with Wrapyfi support☆130Updated 2 years ago