glorgao / SelectiveDPO
☆33Updated last week
Alternatives and similar repositories for SelectiveDPO:
Users that are interested in SelectiveDPO are comparing it to the libraries listed below
- [NeurIPS 2024] "Can Language Models Perform Robust Reasoning in Chain-of-thought Prompting with Noisy Rationales?"☆35Updated 3 months ago
- [ACL'24] Beyond One-Preference-Fits-All Alignment: Multi-Objective Direct Preference Optimization☆74Updated 8 months ago
- Code for paper: Aligning Large Language Models with Representation Editing: A Control Perspective☆28Updated 2 months ago
- Code for "Towards Revealing the Mystery behind Chain of Thought: a Theoretical Perspective"☆20Updated last year
- Official code for "Decoding-Time Language Model Alignment with Multiple Objectives".☆22Updated 5 months ago
- [ACL'24, Outstanding Paper] Emulated Disalignment: Safety Alignment for Large Language Models May Backfire!☆36Updated 8 months ago
- ☆50Updated last year
- Localize-and-Stitch: Efficient Model Merging via Sparse Task Arithmetic☆23Updated 3 months ago
- Lightweight Adapting for Black-Box Large Language Models☆22Updated last year
- ☆39Updated last year
- Code for the ICML 2024 paper "Rewards-in-Context: Multi-objective Alignment of Foundation Models with Dynamic Preference Adjustment"☆64Updated 3 months ago
- Official repository of "Localizing Task Information for Improved Model Merging and Compression" [ICML 2024]☆43Updated 5 months ago
- [ICLR 2025] "Rethinking LLM Unlearning Objectives: A Gradient Perspective and Go Beyond"☆11Updated last month
- [NeurIPS 2023] Github repository for "Composing Parameter-Efficient Modules with Arithmetic Operations"☆60Updated last year
- ☆25Updated 11 months ago
- [ICLR 2025] Unintentional Unalignment: Likelihood Displacement in Direct Preference Optimization☆24Updated 2 months ago
- Official Repository for The Paper: Safety Alignment Should Be Made More Than Just a Few Tokens Deep☆89Updated 9 months ago
- [ICML 2024] Assessing the Brittleness of Safety Alignment via Pruning and Low-Rank Modifications☆76Updated 3 weeks ago
- Awesome-Efficient-Inference-for-LRMs is a collection of state-of-the-art, novel, exciting, token-efficient methods for Large Reasoning Mo…☆56Updated this week
- Direct preference optimization with f-divergences.☆13Updated 5 months ago
- SafeChain: Safety of Language Models with Long Chain-of-Thought Reasoning Capabilities☆12Updated 2 weeks ago
- ☆18Updated last year
- What Makes a Reward Model a Good Teacher? An Optimization Perspective☆23Updated last week
- A curated list of resources for activation engineering☆63Updated 2 weeks ago
- [ACL 2024] Shifting Attention to Relevance: Towards the Predictive Uncertainty Quantification of Free-Form Large Language Models☆47Updated 7 months ago
- ☆90Updated 3 months ago
- Preprint: Asymmetry in Low-Rank Adapters of Foundation Models☆35Updated last year
- [NeurIPS 2023 Spotlight] Temperature Balancing, Layer-wise Weight Analysis, and Neural Network Training☆34Updated 2 weeks ago
- This is the official code for the paper "Safety Tax: Safety Alignment Makes Your Large Reasoning Models Less Reasonable".☆14Updated last month
- [ICLR 2025] Code&Data for the paper "Super(ficial)-alignment: Strong Models May Deceive Weak Models in Weak-to-Strong Generalization"☆13Updated 10 months ago