mallorbc / Finetune_LLMs
Repo for fine-tuning Casual LLMs
☆453Updated 9 months ago
Alternatives and similar repositories for Finetune_LLMs:
Users that are interested in Finetune_LLMs are comparing it to the libraries listed below
- Guide: Finetune GPT2-XL (1.5 Billion Parameters) and finetune GPT-NEO (2.7 B) on a single GPU with Huggingface Transformers using DeepSpe…☆436Updated last year
- Tune any FALCON in 4-bit☆466Updated last year
- ☆536Updated last year
- ☆456Updated last year
- LaMini-LM: A Diverse Herd of Distilled Models from Large-Scale Instructions☆815Updated last year
- This repository contains code for extending the Stanford Alpaca synthetic instruction tuning to existing instruction-tuned models such as…☆350Updated last year
- User-friendly LLaMA: Train or Run the model using PyTorch. Nothing else.☆334Updated last year
- Open-source pre-training implementation of Google's LaMDA in PyTorch. Adding RLHF similar to ChatGPT.☆469Updated 10 months ago
- Alpaca dataset from Stanford, cleaned and curated☆1,531Updated last year
- UI tool for fine-tuning and testing your own LoRA models base on LLaMA, GPT-J and more. One-click run on Google Colab. + A Gradio ChatGPT…☆454Updated last year
- A collection of open-source dataset to train instruction-following LLMs (ChatGPT,LLaMA,Alpaca)☆1,092Updated last year
- Expanding natural instructions☆965Updated last year
- fastLLaMa: An experimental high-performance framework for running Decoder-only LLMs with 4-bit quantization in Python using a C/C++ backe…☆409Updated last year
- Used for adaptive human in the loop evaluation of language and embedding models.☆306Updated last year
- Crosslingual Generalization through Multitask Finetuning☆521Updated 3 months ago
- Ask Me Anything language model prompting☆544Updated last year
- ☆121Updated last year
- Instruct-tune Open LLaMA / RedPajama / StableLM models on consumer hardware using QLoRA☆80Updated last year
- batched loras☆336Updated last year
- Fine-tuning GPT-J-6B on colab or equivalent PC GPU with your custom datasets: 8-bit weights with low-rank adaptors (LoRA)☆74Updated 2 years ago
- simpleT5 is built on top of PyTorch-lightning⚡️ and Transformers🤗 that lets you quickly train your T5 models.☆388Updated last year
- Extend existing LLMs way beyond the original training length with constant memory usage, without retraining☆684Updated 9 months ago
- Finetuning Large Language Models on One Consumer GPU in 2 Bits☆714Updated 7 months ago
- PaL: Program-Aided Language Models (ICML 2023)☆480Updated last year
- Implementation of RETRO, Deepmind's Retrieval based Attention net, in Pytorch☆857Updated last year
- Landmark Attention: Random-Access Infinite Context Length for Transformers☆419Updated last year
- A full pipeline to finetune Vicuna LLM with LoRA and RLHF on consumer hardware. Implementation of RLHF (Reinforcement Learning with Human…☆210Updated 7 months ago
- Code for fine-tuning Platypus fam LLMs using LoRA☆625Updated 11 months ago
- Customizable implementation of the self-instruct paper.☆1,035Updated 10 months ago