A framework for the evaluation of autoregressive code generation language models.
☆1,020Jul 22, 2025Updated 7 months ago
Alternatives and similar repositories for bigcode-evaluation-harness
Users that are interested in bigcode-evaluation-harness are comparing it to the libraries listed below
Sorting:
- A multi-programming language benchmark for LLMs☆298Jan 28, 2026Updated 3 weeks ago
- Rigourous evaluation of LLM-synthesized code - NeurIPS 2023 & COLM 2024☆1,688Oct 2, 2025Updated 4 months ago
- Code for the paper "Evaluating Large Language Models Trained on Code"☆3,137Jan 17, 2025Updated last year
- A framework for few-shot evaluation of language models.☆11,478Feb 15, 2026Updated last week
- 🐙 OctoPack: Instruction Tuning Code Large Language Models☆478Feb 5, 2025Updated last year
- ☆489Aug 15, 2024Updated last year
- CrossCodeEval: A Diverse and Multilingual Benchmark for Cross-File Code Completion (NeurIPS 2023)☆174Aug 15, 2025Updated 6 months ago
- Repository for analysis and experiments in the BigCode project.☆128Mar 20, 2024Updated last year
- [ICML 2023] Data and code release for the paper "DS-1000: A Natural and Reliable Benchmark for Data Science Code Generation".☆266Oct 30, 2024Updated last year
- Run evaluation on LLMs using human-eval benchmark☆427Sep 12, 2023Updated 2 years ago
- AllenAI's post-training codebase☆3,592Updated this week
- [ICLR'25] BigCodeBench: Benchmarking Code Generation Towards AGI☆479Jan 3, 2026Updated last month
- Official repository for the paper "LiveCodeBench: Holistic and Contamination Free Evaluation of Large Language Models for Code"☆803Jul 16, 2025Updated 7 months ago
- ☆1,504May 12, 2023Updated 2 years ago
- SWE-bench: Can Language Models Resolve Real-world Github Issues?☆4,337Feb 19, 2026Updated last week
- Lighteval is your all-in-one toolkit for evaluating LLMs across multiple backends☆2,311Feb 20, 2026Updated last week
- Home of StarCoder: fine-tuning & inference!☆7,530Feb 27, 2024Updated 2 years ago
- Releasing code for "ReCode: Robustness Evaluation of Code Generation Models"☆58Mar 20, 2024Updated last year
- A distributed, extensible, secure solution for evaluating machine generated code with unit tests in multiple programming languages.☆62Oct 21, 2024Updated last year
- Tools for merging pretrained large language models.☆6,814Jan 26, 2026Updated last month
- Code for the paper "Efficient Training of Language Models to Fill in the Middle"☆199Apr 2, 2023Updated 2 years ago
- This is the official code for the paper CodeRL: Mastering Code Generation through Pretrained Models and Deep Reinforcement Learning (Neur…☆558Jan 21, 2025Updated last year
- Beyond the Imitation Game collaborative benchmark for measuring and extrapolating the capabilities of language models☆3,207Jul 19, 2024Updated last year
- Minimalistic large language model 3D-parallelism training☆2,569Feb 19, 2026Updated last week
- Fine-tune SantaCoder for Code/Text Generation.☆196Apr 11, 2023Updated 2 years ago
- ☆85Jun 13, 2023Updated 2 years ago
- ☆672Nov 1, 2024Updated last year
- An automatic evaluator for instruction-following language models. Human-validated, high-quality, cheap, and fast.☆1,953Aug 9, 2025Updated 6 months ago
- [EMNLP'23] Execution-Based Evaluation for Open Domain Code Generation☆49Dec 22, 2023Updated 2 years ago
- LLMs build upon Evol Insturct: WizardLM, WizardCoder, WizardMath☆9,476Jun 7, 2025Updated 8 months ago
- Scaling Data-Constrained Language Models☆340Jun 28, 2025Updated 8 months ago
- ✨ RepoBench: Benchmarking Repository-Level Code Auto-Completion Systems - ICLR 2024☆187Aug 16, 2024Updated last year
- Holistic Evaluation of Language Models (HELM) is an open source Python framework created by the Center for Research on Foundation Models …☆2,684Updated this week
- Robust recipes to align language models with human and AI preferences☆5,506Sep 8, 2025Updated 5 months ago
- 800,000 step-level correctness labels on LLM solutions to MATH problems☆2,091Jun 1, 2023Updated 2 years ago
- Code for the curation of The Stack v2 and StarCoder2 training data☆126Apr 11, 2024Updated last year
- Train transformer language models with reinforcement learning.☆17,460Updated this week
- CRUXEval: Code Reasoning, Understanding, and Execution Evaluation☆166Oct 11, 2024Updated last year
- ☆112Jul 17, 2024Updated last year