Code accompanying the paper Pretraining Language Models with Human Preferences
β180Feb 13, 2024Updated 2 years ago
Alternatives and similar repositories for pretraining-with-human-feedback
Users that are interested in pretraining-with-human-feedback are comparing it to the libraries listed below
Sorting:
- Human preference data for "Training a Helpful and Harmless Assistant with Reinforcement Learning from Human Feedback"β1,818Jun 17, 2025Updated 8 months ago
- Official Code Repository for [AutoScaleπ: Scale-Aware Data Mixing for Pre-Training LLMs] Published as a conference paper at **COLM 2025*β¦β13Aug 8, 2025Updated 7 months ago
- β39Jul 25, 2024Updated last year
- A modular RL library to fine-tune language models to human preferencesβ2,380Mar 1, 2024Updated 2 years ago
- β37May 7, 2023Updated 2 years ago
- β35Jan 29, 2023Updated 3 years ago
- Offline RL experimentsβ15Oct 1, 2022Updated 3 years ago
- β12Jul 8, 2023Updated 2 years ago
- β156Aug 24, 2021Updated 4 years ago
- An original implementation of "MetaICL Learning to Learn In Context" by Sewon Min, Mike Lewis, Luke Zettlemoyer and Hannaneh Hajishirziβ273Apr 15, 2023Updated 2 years ago
- This repository contains some of the code used in the paper "Training Language Models with Langauge Feedback at Scale"β27Mar 30, 2023Updated 2 years ago
- β13Dec 12, 2025Updated 2 months ago
- β58May 30, 2024Updated last year
- A framework for human-readable prompt-based method with large language models. Specially designed for researchers. (Deprecated, check outβ¦β131Feb 25, 2023Updated 3 years ago
- Code for the paper Fine-Tuning Language Models from Human Preferencesβ1,380Jul 25, 2023Updated 2 years ago
- Vision Large Language Models trained on M3IT instruction tuning datasetβ17Aug 16, 2023Updated 2 years ago
- RLHF implementation details of OAI's 2019 codebaseβ197Jan 14, 2024Updated 2 years ago
- Experiments with generating opensource language model assistantsβ97May 14, 2023Updated 2 years ago
- Code for EMNLP 2021 paper: Improving Sequence-to-Sequence Pre-training via Sequence Span Rewritingβ17Nov 30, 2021Updated 4 years ago
- The Intermediate Goal of the project is to train a GPT like architecture to learn to summarise reddit posts from human preferences, as thβ¦β12Jul 14, 2021Updated 4 years ago
- A dataset for realistic evaluation of noisy label methodsβ14Dec 3, 2023Updated 2 years ago
- Code for Contrastive Preference Learning (CPL)β180Nov 22, 2024Updated last year
- Datasets collection and preprocessings framework for NLP extreme multitask learningβ193Jul 9, 2025Updated 8 months ago
- Variational Reinforcement Learningβ17Jul 25, 2024Updated last year
- μ¬μ μμ λν μλ¬Έλ§ μΆμΆν λ°μ΄ν°β16Apr 24, 2023Updated 2 years ago
- This project studies the performance and robustness of language models and task-adaptation methods.β154May 18, 2024Updated last year
- A (somewhat) minimal library for finetuning language models with PPO on human feedback.β90Nov 23, 2022Updated 3 years ago
- Simple next-token-prediction for RLHFβ229Sep 30, 2023Updated 2 years ago
- Keeping language models honest by directly eliciting knowledge encoded in their activations.β217Mar 2, 2026Updated last week
- Sparse Backpropagation for Mixture-of-Expert Trainingβ29Jul 2, 2024Updated last year
- β282Jan 6, 2025Updated last year
- RewardBench: the first evaluation tool for reward models.β697Feb 16, 2026Updated 3 weeks ago
- Convenient Text-to-Text Training for Transformersβ19Dec 10, 2021Updated 4 years ago
- β17May 19, 2023Updated 2 years ago
- β158Mar 18, 2023Updated 2 years ago
- Secrets of RLHF in Large Language Models Part I: PPOβ1,416Mar 3, 2024Updated 2 years ago
- Reproduce results and replicate training fo T0 (Multitask Prompted Training Enables Zero-Shot Task Generalization)β465Nov 5, 2022Updated 3 years ago
- β70Jun 16, 2022Updated 3 years ago
- OpenICL is an open-source framework to facilitate research, development, and prototyping of in-context learning.β584Oct 3, 2023Updated 2 years ago