cambridgeltl / PairSLinks
Aligning with Human Judgement: The Role of Pairwise Preference in Large Language Model Evaluators (Liu et al.; COLM 2024)
☆48Updated last year
Alternatives and similar repositories for PairS
Users that are interested in PairS are comparing it to the libraries listed below
Sorting:
- Repository for NPHardEval, a quantified-dynamic benchmark of LLMs☆63Updated last year
- Scalable Meta-Evaluation of LLMs as Evaluators☆43Updated last year
- ☆75Updated last year
- PASTA: Post-hoc Attention Steering for LLMs☆134Updated last year
- A dataset of LLM-generated chain-of-thought steps annotated with mistake location.☆85Updated last year
- A simple GPT-based evaluation tool for multi-aspect, interpretable assessment of LLMs.☆90Updated 2 years ago
- Code for EMNLP 2024 paper "Learn Beyond The Answer: Training Language Models with Reflection for Mathematical Reasoning"☆54Updated last year
- Codebase for Instruction Following without Instruction Tuning☆36Updated last year
- ☆130Updated last year
- This repository contains data, code and models for contextual noncompliance.☆24Updated last year
- Official code for "MAmmoTH2: Scaling Instructions from the Web" [NeurIPS 2024]☆148Updated last year
- the instructions and demonstrations for building a formal logical reasoning capable GLM☆55Updated last year
- Source code of "Reasons to Reject? Aligning Language Models with Judgments"☆58Updated last year
- [ACL'24] Code and data of paper "When is Tree Search Useful for LLM Planning? It Depends on the Discriminator"☆54Updated last year
- [ICLR 2024] Evaluating Large Language Models at Evaluating Instruction Following☆136Updated last year
- ☆51Updated last year
- Implementation of the paper: "Answering Questions by Meta-Reasoning over Multiple Chains of Thought"☆96Updated 2 years ago
- Fairer Preferences Elicit Improved Human-Aligned Large Language Model Judgments (Zhou et al., EMNLP 2024)☆14Updated last year
- Code for the arXiv preprint "The Unreasonable Effectiveness of Easy Training Data"☆48Updated 2 years ago
- FollowIR: Evaluating and Teaching Information Retrieval Models to Follow Instructions☆51Updated last year
- ☆62Updated 8 months ago
- Co-LLM: Learning to Decode Collaboratively with Multiple Language Models☆125Updated last year
- Official repository for MATES: Model-Aware Data Selection for Efficient Pretraining with Data Influence Models [NeurIPS 2024]☆79Updated last year
- The Official Repository for "Bring Your Own Data! Self-Supervised Evaluation for Large Language Models"☆107Updated 2 years ago
- Evaluate the Quality of Critique☆36Updated last year
- Lightweight tool to identify Data Contamination in LLMs evaluation☆53Updated last year
- Resolving Knowledge Conflicts in Large Language Models, COLM 2024☆18Updated 3 months ago
- ☆161Updated last year
- ☆114Updated 8 months ago
- ☆64Updated last year