Dahoas / reward-modelingView external linksLinks
☆99May 30, 2023Updated 2 years ago
Alternatives and similar repositories for reward-modeling
Users that are interested in reward-modeling are comparing it to the libraries listed below
Sorting:
- A repository for transformer critique learning and generation☆89Dec 7, 2023Updated 2 years ago
- A (somewhat) minimal library for finetuning language models with PPO on human feedback.☆90Nov 23, 2022Updated 3 years ago
- Experiments with generating opensource language model assistants☆97May 14, 2023Updated 2 years ago
- [NIPS2023] RRHF & Wombat☆808Sep 22, 2023Updated 2 years ago
- A repo for distributed training of language models with Reinforcement Learning via Human Feedback (RLHF)☆4,742Jan 8, 2024Updated 2 years ago
- ☆35Jan 29, 2023Updated 3 years ago
- [ICLR 2024] This is the official implementation for the paper: "Beyond imitation: Leveraging fine-grained quality signals for alignment"☆10May 5, 2024Updated last year
- Implementation of ChatGPT RLHF (Reinforcement Learning with Human Feedback) on any generation model in huggingface's transformer (blommz-…☆565May 9, 2024Updated last year
- ☆158Mar 18, 2023Updated 2 years ago
- For experiments involving instruct gpt. Currently used for documenting open research questions.☆71Nov 8, 2022Updated 3 years ago
- ☆12Jan 17, 2025Updated last year
- A modular RL library to fine-tune language models to human preferences☆2,377Mar 1, 2024Updated last year
- ☆33Apr 23, 2023Updated 2 years ago
- A swarm of LLM agents that will help you test, document, and productionize your code!☆16Feb 7, 2026Updated last week
- Training AI for Super Smash Bros. Melee☆32Mar 27, 2025Updated 10 months ago
- Pretraining summarization models using a corpus of nonsense☆13Sep 28, 2021Updated 4 years ago
- ☆14May 8, 2023Updated 2 years ago
- Code for the paper Fine-Tuning Language Models from Human Preferences☆1,377Jul 25, 2023Updated 2 years ago
- Code accompanying the paper Pretraining Language Models with Human Preferences☆180Feb 13, 2024Updated 2 years ago
- Simple next-token-prediction for RLHF☆229Sep 30, 2023Updated 2 years ago
- [ICML 2024] Generalizing Knowledge Graph Embedding with Universal Orthogonal Parameterization☆15May 12, 2024Updated last year
- This is a pip package implementing Reinforcement Learning algorithms in non-stationary environments supported by the OpenAI Gym toolkit.☆16Jun 28, 2024Updated last year
- ☆14Aug 15, 2024Updated last year
- AdamW optimizer for bfloat16 models in pytorch 🔥.☆39Jun 16, 2024Updated last year
- Human preference data for "Training a Helpful and Harmless Assistant with Reinforcement Learning from Human Feedback"☆1,814Jun 17, 2025Updated 7 months ago
- Hidden Engrams: Long Term Memory for Transformer Model Inference☆35Jun 26, 2021Updated 4 years ago
- Pytorch implementation on OpenAI's Procgen ppo-baseline, built from scratch.☆14May 17, 2024Updated last year
- Redwood Research's transformer interpretability tools☆15Apr 15, 2022Updated 3 years ago
- code and data for paper "GIANT: Scalable Creation of a Web-scale Ontology"☆39Apr 22, 2020Updated 5 years ago
- Official code from the paper "Offline RL for Natural Language Generation with Implicit Language Q Learning"☆211Jul 31, 2023Updated 2 years ago
- The code for the video tutorial series on building a Transformer from scratch: https://www.youtube.com/watch?v=XR4VDnJzB8o☆19Apr 15, 2023Updated 2 years ago
- Code for our SIGGRAPH 2023 paper, "Acting as Inverse Inverse Planning"☆19Apr 21, 2023Updated 2 years ago
- Implementation of Reinforcement Learning from Human Feedback (RLHF)☆173Apr 7, 2023Updated 2 years ago
- Reference implementation for Reward-Augmented Decoding: Efficient Controlled Text Generation With a Unidirectional Reward Model☆45Oct 1, 2025Updated 4 months ago
- Demonstration that finetuning RoPE model on larger sequences than the pre-trained model adapts the model context limit☆63Jun 21, 2023Updated 2 years ago
- Text to Speech Synthesis based on controllable latent representation☆14Aug 30, 2019Updated 6 years ago