☆112Nov 7, 2024Updated last year
Alternatives and similar repositories for JudgeBench
Users that are interested in JudgeBench are comparing it to the libraries listed below
Sorting:
- [EMNLP 2025] Dataset and Code of "PersonaGym: Evaluating Persona Agents and LLMs"☆40Aug 21, 2025Updated 7 months ago
- Is In-Context Learning Sufficient for Instruction Following in LLMs? [ICLR 2025]☆32Jan 23, 2025Updated last year
- ☆13Dec 9, 2024Updated last year
- Adversaial attack comparative assessment Large Language Model☆13May 21, 2025Updated 10 months ago
- FBI: Finding Blindspots in LLM Evaluations with Interpretable Checklists☆31Aug 14, 2025Updated 7 months ago
- Arena-Hard-Auto: An automatic LLM benchmark.☆1,008Jun 21, 2025Updated 9 months ago
- ☆19Oct 2, 2023Updated 2 years ago
- ☆535Jul 25, 2025Updated 7 months ago
- [CCS 2024] Optimization-based Prompt Injection Attack to LLM-as-a-Judge☆39Sep 17, 2025Updated 6 months ago
- RewardBench: the first evaluation tool for reward models.☆704Feb 16, 2026Updated last month
- [ICLR 25 Oral] RM-Bench: Benchmarking Reward Models of Language Models with Subtlety and Style☆78Jul 18, 2025Updated 8 months ago
- A universal workflow system for exactly-once DAGs☆23Jun 1, 2023Updated 2 years ago
- Official implementation of "OffsetBias: Leveraging Debiased Data for Tuning Evaluators"☆26Sep 11, 2024Updated last year
- An implementation of online data mixing for the Pile dataset, based on the GPT-NeoX library.☆13Jan 9, 2024Updated 2 years ago
- Systematic evaluation framework that automatically rates overthinking behavior in large language models.☆97May 16, 2025Updated 10 months ago
- Official Implementation of the paper "Jointly Reinforcing Diversity and Quality in Language Model Generations"☆57Dec 26, 2025Updated 2 months ago
- [ICLR 2024] Evaluating Large Language Models at Evaluating Instruction Following☆137Jul 8, 2024Updated last year
- ☆173Oct 12, 2025Updated 5 months ago
- [COLM 2025] Official code for "When To Solve, When To Verify: Compute-Optimal Problem Solving and Generative Verification for LLM Reasoni…☆15Oct 31, 2025Updated 4 months ago
- The Unreliability of Explanations in Few-shot Prompting for Textual Reasoning (NeurIPS 2022)☆16Feb 11, 2023Updated 3 years ago
- [ICLR 2025] LongPO: Long Context Self-Evolution of Large Language Models through Short-to-Long Preference Optimization☆43Feb 27, 2025Updated last year
- Improving Your Model Ranking on Chatbot Arena by Vote Rigging (ICML 2025)☆26Feb 25, 2025Updated last year
- DSTC10 Track 2 - Knowledge-grounded Task-oriented Dialogue Modeling on Spoken Conversations☆62Jul 25, 2023Updated 2 years ago
- An automatic evaluator for instruction-following language models. Human-validated, high-quality, cheap, and fast.☆1,961Aug 9, 2025Updated 7 months ago
- Lightweight PDF Q&A tool powered by RAG (Retrieval-Augmented Generation) with MCP (Model Context Protocol) Support.☆22Oct 27, 2025Updated 4 months ago
- [ACL 2025] Agentic Reward Modeling: Integrating Human Preferences with Verifiable Correctness Signals for Reliable Reward Systems☆125Jun 11, 2025Updated 9 months ago
- To assess the longtext capabilities more comprehensively, we propose Needle-in-a-Haystack PLUS, which shifts the focus from simple fact r…☆13Mar 4, 2024Updated 2 years ago
- Official Code Repository for [AutoScale📈: Scale-Aware Data Mixing for Pre-Training LLMs] Published as a conference paper at **COLM 2025*…☆13Aug 8, 2025Updated 7 months ago
- ☆25May 16, 2024Updated last year
- Pytorch implementation of HyperLLaVA: Dynamic Visual and Language Expert Tuning for Multimodal Large Language Models☆28Mar 22, 2024Updated 2 years ago
- Sotopia-RL: Reward Design for Social Intelligence☆47Jan 29, 2026Updated last month
- ☆32Jun 5, 2025Updated 9 months ago
- ☆34Nov 26, 2025Updated 3 months ago
- ☆63May 13, 2025Updated 10 months ago
- The official evaluation suite and dynamic data release for MixEval.☆255Nov 10, 2024Updated last year
- The implementation of paper "LLM Critics Help Catch Bugs in Mathematics: Towards a Better Mathematical Verifier with Natural Language Fee…☆38Jul 25, 2024Updated last year
- [ICLR 2025 Spotlight] An open-sourced LLM judge for evaluating LLM-generated answers.☆423Feb 11, 2025Updated last year
- Official Github repo for the paper "Evaluating the Evaluation of Diversity in Natural Language Generation"☆21Feb 23, 2021Updated 5 years ago
- ☆16Oct 21, 2024Updated last year