ruizheng20 / gpoLinks
The code of paper "Toward Optimal LLM Alignments Using Two-Player Games".
☆17Updated 11 months ago
Alternatives and similar repositories for gpo
Users that are interested in gpo are comparing it to the libraries listed below
Sorting:
- ☆40Updated last year
- Lightweight Adapting for Black-Box Large Language Models☆22Updated last year
- Domain-specific preference (DSP) data and customized RM fine-tuning.☆25Updated last year
- Source codes for "Preference-grounded Token-level Guidance for Language Model Fine-tuning" (NeurIPS 2023).☆16Updated 4 months ago
- ☆37Updated last year
- ☆41Updated 8 months ago
- ☆49Updated last year
- ☆38Updated 2 months ago
- Restore safety in fine-tuned language models through task arithmetic☆28Updated last year
- Augmenting Statistical Models with Natural Language Parameters☆26Updated 8 months ago
- ☆42Updated last year
- ☆29Updated last year
- ☆18Updated 9 months ago
- Directional Preference Alignment☆56Updated 8 months ago
- [NAACL 2025] The official implementation of paper "Learning From Failure: Integrating Negative Examples when Fine-tuning Large Language M…☆26Updated last year
- Code for ACL 2023 paper "BOLT: Fast Energy-based Controlled Text Generation with Tunable Biases".☆21Updated last year
- Code for NeurIPS 2024 paper "Regularizing Hidden States Enables Learning Generalizable Reward Model for LLMs"☆34Updated 3 months ago
- Code for "Universal Adversarial Triggers Are Not Universal."☆17Updated last year
- A Mechanistic Understanding of Alignment Algorithms: A Case Study on DPO and Toxicity.☆72Updated 2 months ago
- [ACL 2023 Findings] What In-Context Learning “Learns” In-Context: Disentangling Task Recognition and Task Learning☆21Updated last year
- Rewarded soups official implementation☆58Updated last year
- This is a unified platform for implementing and evaluating test-time reasoning mechanisms in Large Language Models (LLMs).☆18Updated 4 months ago
- Official implementation of Bootstrapping Language Models via DPO Implicit Rewards☆44Updated last month
- [ICLR'24] RAIN: Your Language Models Can Align Themselves without Finetuning☆93Updated last year
- This is the official repo for Towards Uncertainty-Aware Language Agent.☆25Updated 9 months ago
- Code for the paper <SelfCheck: Using LLMs to Zero-Shot Check Their Own Step-by-Step Reasoning>☆49Updated last year
- ☆48Updated 3 weeks ago
- Offical code of the paper Large Language Models Are Implicitly Topic Models: Explaining and Finding Good Demonstrations for In-Context Le…☆74Updated last year
- ☆30Updated 11 months ago
- Grade-School Math with Irrelevant Context (GSM-IC) benchmark is an arithmetic reasoning dataset built upon GSM8K, by adding irrelevant se…☆60Updated 2 years ago