The code and data for "MMLU-Pro: A More Robust and Challenging Multi-Task Language Understanding Benchmark" [NeurIPS 2024]
☆352Mar 18, 2026Updated last week
Alternatives and similar repositories for MMLU-Pro
Users that are interested in MMLU-Pro are comparing it to the libraries listed below. We may earn a commission when you buy through links labeled 'Ad' on this page.
Sorting:
- GPQA: A Graduate-Level Google-Proof Q&A Benchmark☆480Sep 30, 2024Updated last year
- Measuring Massive Multitask Language Understanding | ICLR 2021☆1,569May 28, 2023Updated 2 years ago
- The official repo for "TheoremQA: A Theorem-driven Question Answering dataset" (EMNLP 2023)☆38May 15, 2024Updated last year
- ☆110Aug 21, 2025Updated 7 months ago
- The official repository of the Omni-MATH benchmark.☆93Dec 22, 2024Updated last year
- GPU virtual machines on DigitalOcean Gradient AI • AdGet to production fast with high-performance AMD and NVIDIA GPUs you can spin up in seconds. The definition of operational simplicity.
- Official repository for the paper "LiveCodeBench: Holistic and Contamination Free Evaluation of Large Language Models for Code"☆823Jul 16, 2025Updated 8 months ago
- The official repo for “Unleashing the Reasoning Potential of Pre-trained LLMs by Critique Fine-Tuning on One Problem” [EMNLP25]☆34Sep 1, 2025Updated 6 months ago
- A framework for few-shot evaluation of language models.☆11,802Mar 18, 2026Updated last week
- ☆11Jun 11, 2024Updated last year
- ☆31Nov 9, 2024Updated last year
- ☆17Feb 4, 2025Updated last year
- HelloBench: Evaluating Long Text Generation Capabilities of Large Language Models☆54Nov 26, 2024Updated last year
- More reliable Video Understanding Evaluation☆14Sep 23, 2025Updated 6 months ago
- Code for "Critique Fine-Tuning: Learning to Critique is More Effective than Learning to Imitate" [COLM 2025]☆180Jul 8, 2025Updated 8 months ago
- Bare Metal GPUs on DigitalOcean Gradient AI • AdPurpose-built for serious AI teams training foundational models, running large-scale inference, and pushing the boundaries of what's possible.
- RepoQA: Evaluating Long-Context Code Understanding☆130Nov 1, 2024Updated last year
- ☆19Jan 3, 2025Updated last year
- ☆11Oct 11, 2023Updated 2 years ago
- [NAACL'25] "Revealing the Barriers of Language Agents in Planning"☆13Jun 22, 2025Updated 9 months ago
- ☆33Oct 13, 2025Updated 5 months ago
- Modified Beam Search with periodical restart☆12Sep 12, 2024Updated last year
- ☆21Jun 12, 2024Updated last year
- An automatic evaluator for instruction-following language models. Human-validated, high-quality, cheap, and fast.☆1,961Aug 9, 2025Updated 7 months ago
- WMDP is a LLM proxy benchmark for hazardous knowledge in bio, cyber, and chemical security. We also release code for RMU, an unlearning m…☆162May 29, 2025Updated 9 months ago
- Virtual machines for every use case on DigitalOcean • AdGet dependable uptime with 99.99% SLA, simple security tools, and predictable monthly pricing with DigitalOcean's virtual machines, called Droplets.
- The official evaluation suite and dynamic data release for MixEval.☆256Nov 10, 2024Updated last year
- Lighteval is your all-in-one toolkit for evaluating LLMs across multiple backends☆2,353Mar 9, 2026Updated 2 weeks ago
- A simple GPT-based evaluation tool for multi-aspect, interpretable assessment of LLMs.☆90Jan 29, 2024Updated 2 years ago
- LiveBench: A Challenging, Contamination-Free LLM Benchmark☆1,108Updated this week
- [ICLR 2026] Official repository of "InternSVG: Towards Unified SVG Tasks with Multimodal Large Language Models".☆95Feb 6, 2026Updated last month
- [ICLR'25] BigCodeBench: Benchmarking Code Generation Towards AGI☆488Jan 3, 2026Updated 2 months ago
- Preparing for ML Interviews.☆53Jan 12, 2026Updated 2 months ago
- Arena-Hard-Auto: An automatic LLM benchmark.☆1,008Jun 21, 2025Updated 9 months ago
- 800,000 step-level correctness labels on LLM solutions to MATH problems☆2,106Jun 1, 2023Updated 2 years ago
- Simple, predictable pricing with DigitalOcean hosting • AdAlways know what you'll pay with monthly caps and flat pricing. Enterprise-grade infrastructure trusted by 600k+ customers.
- https://x.com/BlinkDL_AI/status/1884768989743882276☆28May 4, 2025Updated 10 months ago
- A light-weight tool for evaluating LLMs in rule-based ways.☆85Jun 19, 2025Updated 9 months ago
- KV Cache Steering for Inducing Reasoning in Small Language Models☆46Jul 24, 2025Updated 8 months ago
- [COLM 2025] Official PyTorch implementation of "Quantization Hurts Reasoning? An Empirical Study on Quantized Reasoning Models"☆72Jul 8, 2025Updated 8 months ago
- Evaluator for LLMs☆27Jan 25, 2024Updated 2 years ago
- ☆4,406Jul 31, 2025Updated 7 months ago
- Nexusflow function call, tool use, and agent benchmarks.☆30Dec 13, 2024Updated last year