princeton-nlp / InstructEval
[NAACL 2024 Findings] Evaluation suite for the systematic evaluation of instruction selection methods.
☆22Updated last year
Alternatives and similar repositories for InstructEval:
Users that are interested in InstructEval are comparing it to the libraries listed below
- [ACL'24] Code and data of paper "When is Tree Search Useful for LLM Planning? It Depends on the Discriminator"☆54Updated last year
- [ICLR'24 spotlight] Tool-Augmented Reward Modeling☆44Updated 2 months ago
- the instructions and demonstrations for building a formal logical reasoning capable GLM☆53Updated 6 months ago
- Domain-specific preference (DSP) data and customized RM fine-tuning.☆24Updated last year
- SLED: Self Logits Evolution Decoding for Improving Factuality in Large Language Model https://arxiv.org/pdf/2411.02433☆23Updated 3 months ago
- ICML 2024 - Official Repository for EXO: Towards Efficient Exact Optimization of Language Model Alignment☆51Updated 8 months ago
- Code for the arXiv preprint "The Unreasonable Effectiveness of Easy Training Data"☆46Updated last year
- Skill-It! A Data-Driven Skills Framework for Understanding and Training Language Models☆43Updated last year
- Code for EMNLP'24 paper - On Diversified Preferences of Large Language Model Alignment☆15Updated 7 months ago
- Evaluate the Quality of Critique☆35Updated 9 months ago
- Code for "Seeking Neural Nuggets: Knowledge Transfer in Large Language Models from a Parametric Perspective"☆32Updated 10 months ago
- Official repository for paper "Weak-to-Strong Extrapolation Expedites Alignment"☆72Updated 9 months ago
- Code for RL4F: Generating Natural Language Feedback with Reinforcement Learning for Repairing Model Outputs. ACL 2023.☆62Updated 3 months ago
- ☆94Updated last year
- The official repository of "Improving Large Language Models via Fine-grained Reinforcement Learning with Minimum Editing Constraint"☆36Updated last year
- This is the oficial repository for "Safer-Instruct: Aligning Language Models with Automated Preference Data"☆17Updated last year
- CodeUltraFeedback: aligning large language models to coding preferences☆69Updated 8 months ago
- [NAACL 2025] A Closer Look into Mixture-of-Experts in Large Language Models☆45Updated last month
- ☆34Updated 11 months ago
- Reference implementation for Reward-Augmented Decoding: Efficient Controlled Text Generation With a Unidirectional Reward Model☆42Updated last year
- Official implementation of paper "Beyond Bradley-Terry Models: A General Preference Model for Language Model Alignment" (https://arxiv.or…☆22Updated 3 weeks ago
- Code and models for EMNLP 2024 paper "WPO: Enhancing RLHF with Weighted Preference Optimization"☆38Updated 5 months ago
- Reproduction of "RLCD Reinforcement Learning from Contrast Distillation for Language Model Alignment☆66Updated last year
- Code for ACL2024 paper - Adversarial Preference Optimization (APO).☆51Updated 9 months ago
- official implementation of paper "Process Reward Model with Q-value Rankings"☆49Updated last month
- Codebase for Instruction Following without Instruction Tuning☆33Updated 5 months ago
- ☆22Updated 5 months ago
- The official implementation of paper "Learning From Failure: Integrating Negative Examples when Fine-tuning Large Language Models as Agen…☆24Updated last year
- Accompanying code for "Boosted Prompt Ensembles for Large Language Models"☆30Updated last year
- Minimal implementation of the Self-Play Fine-Tuning Converts Weak Language Models to Strong Language Models paper (ArXiv 20232401.01335)☆29Updated last year