Shentao-YANG / Preference_Grounded_GuidanceLinks
Source codes for "Preference-grounded Token-level Guidance for Language Model Fine-tuning" (NeurIPS 2023).
☆16Updated 5 months ago
Alternatives and similar repositories for Preference_Grounded_Guidance
Users that are interested in Preference_Grounded_Guidance are comparing it to the libraries listed below
Sorting:
- Domain-specific preference (DSP) data and customized RM fine-tuning.☆25Updated last year
- ☆40Updated last year
- Directional Preference Alignment☆57Updated 9 months ago
- ☆39Updated 3 months ago
- ☆27Updated 2 years ago
- Analyzing LLM Alignment via Token distribution shift☆16Updated last year
- Code for Paper (Preserving Diversity in Supervised Fine-tuning of Large Language Models)☆30Updated last month
- ☆17Updated last year
- ☆18Updated last year
- The Unreliability of Explanations in Few-shot Prompting for Textual Reasoning (NeurIPS 2022)☆15Updated 2 years ago
- Code for ACL2024 paper - Adversarial Preference Optimization (APO).☆54Updated last year
- The code of “Improving Weak-to-Strong Generalization with Scalable Oversight and Ensemble Learning”☆16Updated last year
- ☆41Updated last year
- Code and data used in the paper: "Training on Incorrect Synthetic Data via RL Scales LLM Math Reasoning Eight-Fold"☆30Updated last year
- Self-Supervised Alignment with Mutual Information☆19Updated last year
- The official repository of "Improving Large Language Models via Fine-grained Reinforcement Learning with Minimum Editing Constraint"☆38Updated last year
- BeHonest: Benchmarking Honesty in Large Language Models☆34Updated 10 months ago
- AbstainQA, ACL 2024☆26Updated 8 months ago
- Learning adapter weights from task descriptions☆19Updated last year
- [ACL 2023 Findings] What In-Context Learning “Learns” In-Context: Disentangling Task Recognition and Task Learning☆21Updated last year
- ☆31Updated last year
- Codebase for Hyperdecoders https://arxiv.org/abs/2203.08304☆11Updated 2 years ago
- ☆28Updated last year
- ☆30Updated last year
- ☆29Updated last year
- ☆33Updated last year
- ☆13Updated 7 months ago
- [NAACL 2025] The official implementation of paper "Learning From Failure: Integrating Negative Examples when Fine-tuning Large Language M…☆26Updated last year
- Methods and evaluation for aligning language models temporally☆29Updated last year
- Code for Paper (Policy Optimization in RLHF: The Impact of Out-of-preference Data)☆28Updated last year