Simple next-token-prediction for RLHF
☆229Sep 30, 2023Updated 2 years ago
Alternatives and similar repositories for chain-of-hindsight
Users that are interested in chain-of-hindsight are comparing it to the libraries listed below. We may earn a commission when you buy through links labeled 'Ad' on this page.
Sorting:
- ☆158Mar 18, 2023Updated 3 years ago
- A repo for distributed training of language models with Reinforcement Learning via Human Feedback (RLHF)☆4,745Jan 8, 2024Updated 2 years ago
- ☆284Jan 6, 2025Updated last year
- [NIPS2023] RRHF & Wombat☆808Sep 22, 2023Updated 2 years ago
- A repository for transformer critique learning and generation☆89Dec 7, 2023Updated 2 years ago
- Serverless GPU API endpoints on Runpod - Get Bonus Credits • AdSkip the infrastructure headaches. Auto-scaling, pay-as-you-go, no-ops approach lets you focus on innovating your application.
- ☆75Nov 3, 2023Updated 2 years ago
- A large-scale, fine-grained, diverse preference dataset (and models).☆367Dec 29, 2023Updated 2 years ago
- A simulation framework for RLHF and alternatives. Develop your RLHF method without collecting human data.☆843Jul 1, 2024Updated last year
- Unofficial implementation of Chain of Hindsight (https://arxiv.org/abs/2302.02676) using pytorch and huggingface Trainers.☆11Apr 5, 2023Updated 3 years ago
- ☆72May 22, 2023Updated 2 years ago
- [ICLR 2023] Guess the Instruction! Flipped Learning Makes Language Models Stronger Zero-Shot Learners☆117Jun 28, 2025Updated 10 months ago
- Collection of papers for scalable automated alignment.☆93Oct 22, 2024Updated last year
- Code repository for the c-BTM paper☆109Sep 26, 2023Updated 2 years ago
- Self-Alignment with Principle-Following Reward Models☆170Sep 18, 2025Updated 7 months ago
- Deploy to Railway using AI coding agents - Free Credits Offer • AdUse Claude Code, Codex, OpenCode, and more. Autonomous software development now has the infrastructure to match with Railway.
- Code for the paper "Decomposing the Enigma: Subgoal-based Demonstration Learning for Formal Theorem Proving"☆19May 25, 2023Updated 2 years ago
- ☆11Sep 19, 2025Updated 7 months ago
- (NeurIPS '22) LISA: Learning Interpretable Skill Abstractions - A framework for unsupervised skill learning using Imitation☆29Feb 22, 2023Updated 3 years ago
- ☆314Jun 9, 2024Updated last year
- A repository of projects and datasets under active development by Alignment Lab AI☆22Dec 22, 2023Updated 2 years ago
- Dromedary: towards helpful, ethical and reliable LLMs.☆1,142Sep 18, 2025Updated 7 months ago
- RewardBench: the first evaluation tool for reward models.☆713Feb 16, 2026Updated 2 months ago
- [NeurIPS 2024] OlympicArena: Benchmarking Multi-discipline Cognitive Reasoning for Superintelligent AI☆106Mar 6, 2025Updated last year
- Multi-agent Social Simulation + Efficient, Effective, and Stable alternative of RLHF. Code for the paper "Training Socially Aligned Langu…☆355Jun 18, 2023Updated 2 years ago
- Deploy to Railway using AI coding agents - Free Credits Offer • AdUse Claude Code, Codex, OpenCode, and more. Autonomous software development now has the infrastructure to match with Railway.
- Secrets of RLHF in Large Language Models Part I: PPO☆1,424Mar 3, 2024Updated 2 years ago
- [ICML 2023] Exploring the Benefits of Training Expert Language Models over Instruction Tuning☆99Apr 26, 2023Updated 3 years ago
- ☆260Dec 21, 2022Updated 3 years ago
- Self-Supervised Alignment with Mutual Information☆20May 24, 2024Updated last year
- Scaling Data-Constrained Language Models☆343Jun 28, 2025Updated 10 months ago
- Code accompanying the paper Pretraining Language Models with Human Preferences☆181Feb 13, 2024Updated 2 years ago
- A library with extensible implementations of DPO, KTO, PPO, ORPO, and other human-aware loss functions (HALOs).☆904Sep 30, 2025Updated 7 months ago
- [ACL 2023] Knowledge Unlearning for Mitigating Privacy Risks in Language Models☆88Sep 12, 2024Updated last year
- Safe RLHF: Constrained Value Alignment via Safe Reinforcement Learning from Human Feedback☆1,599Nov 24, 2025Updated 5 months ago
- Managed Kubernetes at scale on DigitalOcean • AdDigitalOcean Kubernetes includes the control plane, bandwidth allowance, container registry, automatic updates, and more for free.
- 800,000 step-level correctness labels on LLM solutions to MATH problems☆2,119Jun 1, 2023Updated 2 years ago
- Accompanying repo for the RLPrompt paper☆363Jun 6, 2024Updated last year
- All-in-one repository for Fine-tuning & Pretraining (Large) Language Models☆15Mar 8, 2023Updated 3 years ago
- TART: A plug-and-play Transformer module for task-agnostic reasoning☆202Jun 22, 2023Updated 2 years ago
- Code for Arxiv 2023: Improving Language Model Negociation with Self-Play and In-Context Learning from AI Feedback☆208May 24, 2023Updated 2 years ago
- A modular RL library to fine-tune language models to human preferences☆2,388Mar 1, 2024Updated 2 years ago
- [EMNLP 2023] The CoT Collection: Improving Zero-shot and Few-shot Learning of Language Models via Chain-of-Thought Fine-Tuning☆256Oct 31, 2023Updated 2 years ago