junkangwu / Dr_DPO
[ICLR 2025] Official code of "Towards Robust Alignment of Language Models: Distributionally Robustifying Direct Preference Optimization"
☆11Updated 10 months ago
Alternatives and similar repositories for Dr_DPO:
Users that are interested in Dr_DPO are comparing it to the libraries listed below
- ☆22Updated last week
- [ACL'24] Beyond One-Preference-Fits-All Alignment: Multi-Objective Direct Preference Optimization☆72Updated 7 months ago
- [NeurIPS 2024] Official code of $\beta$-DPO: Direct Preference Optimization with Dynamic $\beta$☆41Updated 5 months ago
- Official implementation for "ALI-Agent: Assessing LLMs'Alignment with Human Values via Agent-based Evaluation"☆16Updated last month
- ☆18Updated 5 months ago
- Code for the ICML 2024 paper "Rewards-in-Context: Multi-objective Alignment of Foundation Models with Dynamic Preference Adjustment"☆62Updated 3 months ago
- [NeurIPS2023] Official code of "Understanding Contrastive Learning via Distributionally Robust Optimization"☆39Updated last year
- Code for "Neural Retrievers are Biased Towards LLM-Generated Content"☆14Updated 5 months ago
- This is a unified platform for implementing and evaluating test-time reasoning mechanisms in Large Language Models (LLMs).☆15Updated 2 months ago
- official implementation of ICLR'2025 paper: Rethinking Bradley-Terry Models in Preference-based Reward Modeling: Foundations, Theory, and…☆37Updated 3 weeks ago
- Official code for "Decoding-Time Language Model Alignment with Multiple Objectives".☆19Updated 5 months ago
- This is the repo for the survey of Bias and Fairness in IR with LLMs.☆52Updated 3 weeks ago
- ☆25Updated 10 months ago
- Direct preference optimization with f-divergences.☆13Updated 5 months ago
- The official repository of "SmartAgent: Chain-of-User-Thought for Embodied Personalized Agent in Cyber World".☆25Updated 2 weeks ago
- ☆82Updated 3 months ago
- ☆15Updated this week
- Repo of "Large Language Model-based Human-Agent Collaboration for Complex Task Solving(EMNLP2024 Findings)"☆31Updated 6 months ago
- Official Implementation of "Democratizing Large Language Models via Personalized Parameter-Efficient Fine-tuning" at EMNLP 2024 Main Conf…☆26Updated 2 months ago
- ☆43Updated 5 months ago
- [NAACL 25 main] Awesome LLM Causal Reasoning is a collection of LLM-based casual reasoning works, including papers, codes and datasets.☆49Updated last month
- ☆18Updated 5 months ago
- ☆19Updated last month
- The code of paper "Toward Optimal LLM Alignments Using Two-Player Games".☆16Updated 9 months ago
- SafeChain: Safety of Language Models with Long Chain-of-Thought Reasoning Capabilities☆12Updated last week
- Code for paper: Aligning Large Language Models with Representation Editing: A Control Perspective☆25Updated 2 months ago
- What Makes a Reward Model a Good Teacher? An Optimization Perspective☆15Updated last week
- ☆37Updated last year
- code for paper Query-Dependent Prompt Evaluation and Optimization with Offline Inverse Reinforcement Learning☆39Updated last year
- An implementation of SEAL: Safety-Enhanced Aligned LLM fine-tuning via bilevel data selection.☆12Updated last month