mengdi-li / awesome-RLAIFLinks
A continually updated list of literature on Reinforcement Learning from AI Feedback (RLAIF)
☆177Updated last week
Alternatives and similar repositories for awesome-RLAIF
Users that are interested in awesome-RLAIF are comparing it to the libraries listed below
Sorting:
- AdaPlanner: Language Models for Decision Making via Adaptive Planning from Feedback☆111Updated 4 months ago
- A brief and partial summary of RLHF algorithms.☆131Updated 5 months ago
- Code for Paper (ReMax: A Simple, Efficient and Effective Reinforcement Learning Method for Aligning Large Language Models)☆189Updated last year
- Reference implementation for Token-level Direct Preference Optimization(TDPO)☆143Updated 5 months ago
- Trial and Error: Exploration-Based Trajectory Optimization of LLM Agents (ACL 2024 Main Conference)☆147Updated 9 months ago
- [NeurIPS 2024 Oral] Aligner: Efficient Alignment by Learning to Correct☆181Updated 6 months ago
- AI Alignment: A Comprehensive Survey☆135Updated last year
- Research Code for "ArCHer: Training Language Model Agents via Hierarchical Multi-Turn RL"☆185Updated 3 months ago
- ☆114Updated 6 months ago
- Paper collections of the continuous effort start from World Models.☆179Updated last year
- ☆203Updated 4 months ago
- Reasoning with Language Model is Planning with World Model☆168Updated last year
- Code for Paper: Autonomous Evaluation and Refinement of Digital Agents [COLM 2024]☆139Updated 8 months ago
- An extensible benchmark for evaluating large language models on planning☆393Updated last month
- Implementation of the ICML 2024 paper "Training Large Language Models for Reasoning through Reverse Curriculum Reinforcement Learning" pr…☆107Updated last year
- ☆152Updated 7 months ago
- [NeurIPS 2024] Agent Planning with World Knowledge Model☆144Updated 7 months ago
- Reproduction of "RLCD Reinforcement Learning from Contrast Distillation for Language Model Alignment☆69Updated last year
- Code and example data for the paper: Rule Based Rewards for Language Model Safety☆190Updated last year
- [NeurIPS 2024] The official implementation of paper: Chain of Preference Optimization: Improving Chain-of-Thought Reasoning in LLMs.☆125Updated 4 months ago
- An index of algorithms for reinforcement learning from human feedback (rlhf))☆92Updated last year
- ☆43Updated 5 months ago
- Direct Preference Optimization from scratch in PyTorch☆103Updated 4 months ago
- Curation of resources for LLM mathematical reasoning, most of which are screened by @tongyx361 to ensure high quality and accompanied wit…☆133Updated last year
- Sotopia-π: Interactive Learning of Socially Intelligent Language Agents (ACL 2024)☆70Updated last year
- Must-read Papers on Large Language Model (LLM) Planning.☆423Updated last year
- [NeurIPS 2023] We use large language models as commonsense world model and heuristic policy within Monte-Carlo Tree Search, enabling bett…☆281Updated 8 months ago
- ☆117Updated 4 months ago
- This is the repository that contains the source code for the Self-Evaluation Guided MCTS for online DPO.☆319Updated last year
- augmented LLM with self reflection☆129Updated last year