arjunbansal / awesome-oss-llm-ift-rlhf
Collection of open source implementations of LLMs with IFT and RLHF that are striving to get to ChatGPT level of performance
β51Updated last year
Alternatives and similar repositories for awesome-oss-llm-ift-rlhf:
Users that are interested in awesome-oss-llm-ift-rlhf are comparing it to the libraries listed below
- Comprehensive analysis of difference in performance of QLora, Lora, and Full Finetunes.β82Updated last year
- Lightweight demos for finetuning LLMs. Powered by π€ transformers and open-source datasets.β66Updated 3 months ago
- Patch for MPT-7B which allows using and training a LoRAβ58Updated last year
- Reimplementation of the task generation part from the Alpaca paperβ119Updated last year
- Multi-Domain Expert Learningβ67Updated last year
- QLoRA: Efficient Finetuning of Quantized LLMsβ77Updated 9 months ago
- Zeus LLM Trainer is a rewrite of Stanford Alpaca aiming to be the trainer for all Large Language Modelsβ69Updated last year
- Code repository for the c-BTM paperβ105Updated last year
- β94Updated last year
- LLM finetuningβ42Updated last year
- Official code for ACL 2023 (short, findings) paper "Recursion of Thought: A Divide and Conquer Approach to Multi-Context Reasoning with Lβ¦β42Updated last year
- Experiments with generating opensource language model assistantsβ97Updated last year
- QLoRA with Enhanced Multi GPU Supportβ36Updated last year
- Small and Efficient Mathematical Reasoning LLMsβ71Updated last year
- Finetune Falcon, LLaMA, MPT, and RedPajama on consumer hardware using PEFT LoRAβ101Updated 6 months ago
- The Next Generation Multi-Modality Superintelligenceβ70Updated 4 months ago
- β74Updated last year
- β24Updated last year
- This repository contains code for cleaning your training data of benchmark data to help combat data snooping.β25Updated last year
- Exploring finetuning public checkpoints on filter 8K sequences on Pileβ115Updated last year
- β84Updated last year
- ToK aka Tree of Knowledge for Large Language Models LLM. It's a novel dataset that inspires knowledge symbolic correlation in simple inpuβ¦β50Updated last year
- Pre-training code for CrystalCoder 7B LLMβ55Updated 8 months ago
- TART: A plug-and-play Transformer module for task-agnostic reasoningβ194Updated last year
- Modified Stanford-Alpaca Trainer for Training Replit's Code Modelβ40Updated last year
- β37Updated last year
- Demonstration that finetuning RoPE model on larger sequences than the pre-trained model adapts the model context limitβ63Updated last year
- inference code for mixtral-8x7b-32kseqlenβ99Updated last year
- Evaluating LLMs with CommonGen-Liteβ88Updated 10 months ago
- Code for NeurIPS LLM Efficiency Challengeβ54Updated 9 months ago