rasbt / low-rank-adaptation-blogLinks
☆29Updated 2 years ago
Alternatives and similar repositories for low-rank-adaptation-blog
Users that are interested in low-rank-adaptation-blog are comparing it to the libraries listed below
Sorting:
- Reward Model framework for LLM RLHF☆62Updated 2 years ago
- Lightweight demos for finetuning LLMs. Powered by 🤗 transformers and open-source datasets.☆78Updated last year
- This is the repo for the paper Shepherd -- A Critic for Language Model Generation☆221Updated 2 years ago
- This project studies the performance and robustness of language models and task-adaptation methods.☆155Updated last year
- Open Implementations of LLM Analyses☆107Updated last year
- ☆78Updated 2 years ago
- Notus is a collection of fine-tuned LLMs using SFT, DPO, SFT+DPO, and/or any other RLHF techniques, while always keeping a data-first app…☆169Updated last year
- ☆86Updated last year
- Data preparation code for Amber 7B LLM☆94Updated last year
- Mixing Language Models with Self-Verification and Meta-Verification☆111Updated last year
- Python tools for processing the stackexchange data dumps into a text dataset for Language Models☆86Updated 2 years ago
- ☆85Updated 2 years ago
- Small and Efficient Mathematical Reasoning LLMs☆73Updated last year
- Exploring finetuning public checkpoints on filter 8K sequences on Pile☆116Updated 2 years ago
- Code for NeurIPS LLM Efficiency Challenge☆59Updated last year
- experiments with inference on llama☆103Updated last year
- ModuleFormer is a MoE-based architecture that includes two different types of experts: stick-breaking attention heads and feedforward exp…☆226Updated 3 months ago
- Pre-training code for Amber 7B LLM☆170Updated last year
- Codebase accompanying the Summary of a Haystack paper.☆80Updated last year
- ☆162Updated last year
- The data processing pipeline for the Koala chatbot language model☆118Updated 2 years ago
- Code accompanying the paper Pretraining Language Models with Human Preferences☆180Updated last year
- Model, Code & Data for the EMNLP'23 paper "Making Large Language Models Better Data Creators"☆137Updated 2 years ago
- ☆180Updated 2 years ago
- Scaling Data-Constrained Language Models☆343Updated 6 months ago
- Finetune Falcon, LLaMA, MPT, and RedPajama on consumer hardware using PEFT LoRA☆104Updated 7 months ago
- Multipack distributed sampler for fast padding-free training of LLMs☆203Updated last year
- Inference script for Meta's LLaMA models using Hugging Face wrapper☆110Updated 2 years ago
- Open Instruction Generalist is an assistant trained on massive synthetic instructions to perform many millions of tasks☆210Updated last year
- Adversarial Training and SFT for Bot Safety Models☆40Updated 2 years ago