OSU-NLP-Group / In-Context-RerankingLinks
[ICLR'25] "Attention in Large Language Models Yields Efficient Zero-Shot Re-Rankers"
☆35Updated 6 months ago
Alternatives and similar repositories for In-Context-Reranking
Users that are interested in In-Context-Reranking are comparing it to the libraries listed below
Sorting:
- ☆74Updated last year
- FollowIR: Evaluating and Teaching Information Retrieval Models to Follow Instructions☆47Updated last year
- [ICLR'24 spotlight] Tool-Augmented Reward Modeling☆51Updated 4 months ago
- Official code for "MAmmoTH2: Scaling Instructions from the Web" [NeurIPS 2024]☆148Updated 11 months ago
- Benchmarking Benchmark Leakage in Large Language Models☆55Updated last year
- ☆47Updated 6 months ago
- [NeurIPS 2024] Train LLMs with diverse system messages reflecting individualized preferences to generalize to unseen system messages☆50Updated last month
- ☆48Updated last year
- [ACL'24] Code and data of paper "When is Tree Search Useful for LLM Planning? It Depends on the Discriminator"☆54Updated last year
- Neuro-Symbolic Integration Brings Causal and Reliable Reasoning Proofs☆39Updated last year
- [ICLR 2025] BRIGHT: A Realistic and Challenging Benchmark for Reasoning-Intensive Retrieval☆167Updated 3 weeks ago
- A dataset of LLM-generated chain-of-thought steps annotated with mistake location.☆82Updated last year
- the instructions and demonstrations for building a formal logical reasoning capable GLM☆54Updated last year
- [ICLR'25] Data and code for our paper "Why Does the Effective Context Length of LLMs Fall Short?"☆77Updated 10 months ago
- Co-LLM: Learning to Decode Collaboratively with Multiple Language Models☆121Updated last year
- Evaluate the Quality of Critique☆36Updated last year
- ☆52Updated last year
- ☆35Updated last year
- Scalable Meta-Evaluation of LLMs as Evaluators☆42Updated last year
- [COLM 2025] EvalTree: Profiling Language Model Weaknesses via Hierarchical Capability Trees☆24Updated 2 months ago
- Code for EMNLP 2024 paper "Learn Beyond The Answer: Training Language Models with Reflection for Mathematical Reasoning"☆55Updated last year
- Source code of "Reasons to Reject? Aligning Language Models with Judgments"☆58Updated last year
- [NAACL 2024] Struc-Bench: Are Large Language Models Good at Generating Complex Structured Tabular Data? https://aclanthology.org/2024.naa…☆55Updated 2 months ago
- Codebase for Instruction Following without Instruction Tuning☆35Updated last year
- Aligning with Human Judgement: The Role of Pairwise Preference in Large Language Model Evaluators (Liu et al.; COLM 2024)☆48Updated 8 months ago
- [EMNLP 2024] A Retrieval Benchmark for Scientific Literature Search☆98Updated 10 months ago
- Exploration of automated dataset selection approaches at large scales.☆47Updated 7 months ago
- This repository includes a benchmark and code for the paper "Evaluating LLMs at Detecting Errors in LLM Responses".☆30Updated last year
- [ACL 2025 Main] Official Repository for "Evaluating Language Models as Synthetic Data Generators"☆38Updated 9 months ago
- Middleware for LLMs: Tools Are Instrumental for Language Agents in Complex Environments (EMNLP'2024)☆37Updated 9 months ago