vicgalle / refined-dpoLinks
Refined Direct Preference Optimization with Synthetic Data for Behavioral Alignment of LLMs
☆13Updated last year
Alternatives and similar repositories for refined-dpo
Users that are interested in refined-dpo are comparing it to the libraries listed below
Sorting:
- Script for processing OpenAI's PRM800K process supervision dataset into an Alpaca-style instruction-response format☆27Updated last year
- Anchored Preference Optimization and Contrastive Revisions: Addressing Underspecification in Alignment☆57Updated 9 months ago
- [ACL 2025] Are Your LLMs Capable of Stable Reasoning?☆25Updated 2 months ago
- ReBase: Training Task Experts through Retrieval Based Distillation☆29Updated 4 months ago
- Scalable Meta-Evaluation of LLMs as Evaluators☆42Updated last year
- Aioli: A unified optimization framework for language model data mixing☆27Updated 4 months ago
- Verifiers for LLM Reinforcement Learning☆56Updated last month
- Codebase for Instruction Following without Instruction Tuning☆34Updated 8 months ago
- ☆22Updated 5 months ago
- ☆27Updated 9 months ago
- Official repository of "LiNeS: Post-training Layer Scaling Prevents Forgetting and Enhances Model Merging"☆26Updated 7 months ago
- ☆24Updated 8 months ago
- Implementation of the model: "Reka Core, Flash, and Edge: A Series of Powerful Multimodal Language Models" in PyTorch☆29Updated last week
- ☆24Updated 4 months ago
- Reference implementation for Reward-Augmented Decoding: Efficient Controlled Text Generation With a Unidirectional Reward Model☆43Updated last year
- Are LLMs Capable of Data-based Statistical and Causal Reasoning? Benchmarking Advanced Quantitative Reasoning with Data☆37Updated 3 months ago
- EMNLP 2024 "Re-reading improves reasoning in large language models". Simply repeating the question to get bidirectional understanding for…☆26Updated 5 months ago
- ZYN: Zero-Shot Reward Models with Yes-No Questions☆34Updated last year
- The repository contains code for Adaptive Data Optimization☆24Updated 5 months ago
- In-Context Alignment: Chat with Vanilla Language Models Before Fine-Tuning☆35Updated last year
- Code for preprint "Metadata Conditioning Accelerates Language Model Pre-training (MeCo)"☆39Updated last month
- Byte-sized text games for code generation tasks on virtual environments☆19Updated 11 months ago
- ☆15Updated 2 months ago
- ☆27Updated this week
- Aligning with Human Judgement: The Role of Pairwise Preference in Large Language Model Evaluators (Liu et al.; COLM 2024)☆47Updated 4 months ago
- Introducing Filtered Direct Preference Optimization (fDPO) that enhances language model alignment with human preferences by discarding lo…☆15Updated 6 months ago
- Tasks for describing differences between text distributions.☆16Updated 9 months ago
- This is official project in our paper: Is Bigger and Deeper Always Better? Probing LLaMA Across Scales and Layers☆30Updated last year
- ☆25Updated 2 years ago
- ☆69Updated last year