Evaluation suite for LLMs
☆379Jul 11, 2025Updated 7 months ago
Alternatives and similar repositories for OLMo-Eval-Legacy
Users that are interested in OLMo-Eval-Legacy are comparing it to the libraries listed below
Sorting:
- Modeling, training, eval, and inference code for OLMo☆6,353Nov 24, 2025Updated 3 months ago
- This project studies the performance and robustness of language models and task-adaptation methods.☆154May 18, 2024Updated last year
- Data and tools for generating and inspecting OLMo pre-training data.☆1,434Nov 5, 2025Updated 4 months ago
- AllenAI's post-training codebase☆3,614Updated this week
- Reproducible, flexible LLM evaluations☆348Mar 2, 2026Updated last week
- Train Models Contrastively in Pytorch☆777Mar 26, 2025Updated 11 months ago
- Gantry is a CLI that streamlines running experiments in Beaker☆32Updated this week
- Freeing data processing from scripting madness by providing a set of platform-agnostic customizable pipeline processing blocks.☆2,915Updated this week
- What's In My Big Data (WIMBD) - a toolkit for analyzing large text datasets☆226Nov 16, 2024Updated last year
- A framework for few-shot evaluation of language models.☆11,618Updated this week
- Scalable Meta-Evaluation of LLMs as Evaluators☆43Feb 15, 2024Updated 2 years ago
- The official evaluation suite and dynamic data release for MixEval.☆255Nov 10, 2024Updated last year
- Robust recipes to align language models with human and AI preferences☆5,510Sep 8, 2025Updated 6 months ago
- OLMoE: Open Mixture-of-Experts Language Models☆982Sep 23, 2025Updated 5 months ago
- Minimalistic large language model 3D-parallelism training☆2,588Feb 19, 2026Updated 2 weeks ago
- Tools for merging pretrained large language models.☆6,842Feb 28, 2026Updated last week
- Lighteval is your all-in-one toolkit for evaluating LLMs across multiple backends☆2,324Updated this week
- [ICLR 2025 Spotlight] An open-sourced LLM judge for evaluating LLM-generated answers.☆420Feb 11, 2025Updated last year
- Repository for NPHardEval, a quantified-dynamic benchmark of LLMs☆63Mar 26, 2024Updated last year
- Lightweight tool to identify Data Contamination in LLMs evaluation☆53Mar 8, 2024Updated 2 years ago
- ☆27Mar 21, 2024Updated last year
- The hub for EleutherAI's work on interpretability and learning dynamics☆2,743Nov 15, 2025Updated 3 months ago
- [ICML 2024] LLMCompiler: An LLM Compiler for Parallel Function Calling☆1,827Jul 10, 2024Updated last year
- Manage scalable open LLM inference endpoints in Slurm clusters☆282Jul 11, 2024Updated last year
- DataComp for Language Models☆1,425Sep 9, 2025Updated 6 months ago
- Distilabel is a framework for synthetic data and AI feedback for engineers who need fast, reliable and scalable pipelines based on verifi…☆3,114Mar 2, 2026Updated last week
- A unified evaluation framework for large language models☆2,780Feb 20, 2026Updated 2 weeks ago
- OpenCompass is an LLM evaluation platform, supporting a wide range of models (Llama3, Mistral, InternLM2,GPT-4,LLaMa2, Qwen,GLM, Claude, …☆6,705Feb 27, 2026Updated last week
- An automatic evaluator for instruction-following language models. Human-validated, high-quality, cheap, and fast.☆1,953Aug 9, 2025Updated 7 months ago
- Skywork-MoE: A Deep Dive into Training Techniques for Mixture-of-Experts Language Models☆139Jun 12, 2024Updated last year
- Human preference data for "Training a Helpful and Harmless Assistant with Reinforcement Learning from Human Feedback"☆1,824Jun 17, 2025Updated 8 months ago
- Holistic Evaluation of Language Models (HELM) is an open source Python framework created by the Center for Research on Foundation Models …☆2,702Updated this week
- Data mapping framework for rust stuff☆47Updated this week
- Website for hosting the Open Foundation Models Cheat Sheet.☆270May 7, 2025Updated 10 months ago
- Open Implementations of LLM Analyses☆107Oct 8, 2024Updated last year
- 【TMM 2025🔥】 Mixture-of-Experts for Large Vision-Language Models☆2,303Jul 15, 2025Updated 7 months ago
- Viewer for text datasets in formats like HuggingFace, JSONL, etc.☆15Feb 25, 2025Updated last year
- Automatically evaluate your LLMs in Google Colab☆687May 7, 2024Updated last year
- Run safety benchmarks against AI models and view detailed reports showing how well they performed.☆120Mar 3, 2026Updated last week