THUDM / BattleAgentBenchLinks
☆3Updated 3 months ago
Alternatives and similar repositories for BattleAgentBench
Users that are interested in BattleAgentBench are comparing it to the libraries listed below
Sorting:
- Rewarded soups official implementation☆58Updated last year
- Direct preference optimization with f-divergences.☆13Updated 7 months ago
- Official code for "Decoding-Time Language Model Alignment with Multiple Objectives".☆23Updated 7 months ago
- A repo for RLHF training and BoN over LLMs, with support for reward model ensembles.☆43Updated 4 months ago
- code for paper Query-Dependent Prompt Evaluation and Optimization with Offline Inverse Reinforcement Learning☆41Updated last year
- Code for NeurIPS 2024 paper "Regularizing Hidden States Enables Learning Generalizable Reward Model for LLMs"☆34Updated 3 months ago
- [ACL'24, Outstanding Paper] Emulated Disalignment: Safety Alignment for Large Language Models May Backfire!☆36Updated 10 months ago
- Tracking literature and additional online resources on transformers for sequential decision making including RL and beyond.☆47Updated 2 years ago
- Code repository for the NAACL 2025 paper "LLM-Coordination: Evaluating and Analyzing Multi-agent Coordination Abilities in Large Language…☆36Updated 7 months ago
- ☆40Updated last year
- Official code for paper Understanding the Reasoning Ability of Language Models From the Perspective of Reasoning Paths Aggregation☆20Updated last year
- Code for the ICML 2024 paper "Rewards-in-Context: Multi-objective Alignment of Foundation Models with Dynamic Preference Adjustment"☆69Updated 5 months ago
- Easy-to-Hard Generalization: Scalable Alignment Beyond Human Supervision☆120Updated 8 months ago
- ☆93Updated 11 months ago
- ☆11Updated 5 months ago
- Code and data for the paper "Understanding Hidden Context in Preference Learning: Consequences for RLHF"☆29Updated last year
- ☆44Updated 2 years ago
- [ICML 2024] Language Models Represent Beliefs of Self and Others☆32Updated 8 months ago
- Code release for "Generating Code World Models with Large Language Models Guided by Monte Carlo Tree Search" published at NeurIPS '24.☆11Updated 3 months ago
- Source codes for "Preference-grounded Token-level Guidance for Language Model Fine-tuning" (NeurIPS 2023).☆16Updated 5 months ago
- Code for our paper LLaMAR: LM-based Long-Horizon Planner for Multi-Agent Robotics☆13Updated 3 months ago
- This is an official implementation of the paper ``Building Math Agents with Multi-Turn Iterative Preference Learning'' with multi-turn DP…☆25Updated 6 months ago
- (AAAI24 oral) Implementation of RPPO(Risk-sensitive PPO) and RPBT(Population-based self-play with RPPO)☆12Updated 2 years ago
- General-Sum variant of the game Diplomacy for evaluating AIs.☆29Updated last year
- ☆16Updated 6 months ago
- Code for Paper (Policy Optimization in RLHF: The Impact of Out-of-preference Data)☆28Updated last year
- ☆15Updated 7 months ago
- [ACL'24] Beyond One-Preference-Fits-All Alignment: Multi-Objective Direct Preference Optimization☆79Updated 9 months ago
- ☆14Updated 7 months ago
- What Makes a Reward Model a Good Teacher? An Optimization Perspective☆31Updated last month