microsoft / prose-benchmarksLinks
PROSE Public Benchmark Suite
☆27Updated last month
Alternatives and similar repositories for prose-benchmarks
Users that are interested in prose-benchmarks are comparing it to the libraries listed below
Sorting:
- Astraios: Parameter-Efficient Instruction Tuning Code Language Models☆62Updated last year
- xCodeEval: A Large Scale Multilingual Multitask Benchmark for Code Understanding, Generation, Translation and Retrieval☆86Updated 11 months ago
- NaturalCodeBench (Findings of ACL 2024)☆67Updated 10 months ago
- [EMNLP'23] Execution-Based Evaluation for Open Domain Code Generation☆48Updated last year
- InstructCoder: Instruction Tuning Large Language Models for Code Editing | Oral ACL-2024 srw☆62Updated 11 months ago
- ☆78Updated 5 months ago
- ☆52Updated last year
- [NAACL 2024] Struc-Bench: Are Large Language Models Good at Generating Complex Structured Tabular Data? https://aclanthology.org/2024.naa…☆54Updated last month
- ☆38Updated 2 months ago
- Official repo for NAACL 2024 Findings paper "LeTI: Learning to Generate from Textual Interactions."☆64Updated 2 years ago
- Language Models of Code are Few-Shot Commonsense Learners (EMNLP 2022)☆86Updated 2 years ago
- CodeUltraFeedback: aligning large language models to coding preferences (TOSEM 2025)☆72Updated last year
- LMTuner: Make the LLM Better for Everyone☆36Updated last year
- ☆20Updated 4 months ago
- LLMs as Collaboratively Edited Knowledge Bases☆45Updated last year
- Plug in and play implementation of " Textbooks Are All You Need", ready for training, inference, and dataset generation☆75Updated last year
- StepCoder: Improve Code Generation with Reinforcement Learning from Compiler Feedback☆68Updated last year
- Implementation of "LM-Infinite: Simple On-the-Fly Length Generalization for Large Language Models"☆40Updated 9 months ago
- [EMNLP 2023 Industry Track] A simple prompting approach that enables the LLMs to run inference in batches.☆76Updated last year
- [NAACL 2024] Enhancing Chain-of-Thoughts Prompting with Iterative Bootstrapping in Large Language Models☆85Updated last year
- Accepted by Transactions on Machine Learning Research (TMLR)☆130Updated 10 months ago
- A dataset of LLM-generated chain-of-thought steps annotated with mistake location.☆81Updated last year
- ☆70Updated last year
- RepoQA: Evaluating Long-Context Code Understanding☆115Updated 10 months ago
- ☆28Updated this week
- Implementation of the model: "Reka Core, Flash, and Edge: A Series of Powerful Multimodal Language Models" in PyTorch☆29Updated 3 weeks ago
- A simple GPT-based evaluation tool for multi-aspect, interpretable assessment of LLMs.☆87Updated last year
- Data preparation code for CrystalCoder 7B LLM☆45Updated last year
- [NeurIPS 2024] Evaluation harness for SWT-Bench, a benchmark for evaluating LLM repository-level test-generation☆54Updated last week
- ☆28Updated 7 months ago