Evaluate your LLM's response with Prometheus and GPT4 💯
☆1,057Apr 25, 2025Updated 10 months ago
Alternatives and similar repositories for prometheus-eval
Users that are interested in prometheus-eval are comparing it to the libraries listed below
Sorting:
- [ICLR 2024 & NeurIPS 2023 WS] An Evaluator LM that is open-source, offers reproducible evaluation, and inexpensive to use. Specifically d…☆312Nov 11, 2023Updated 2 years ago
- [ICLR 2024 Spotlight] FLASK: Fine-grained Language Model Evaluation based on Alignment Skill Sets☆217Dec 24, 2023Updated 2 years ago
- [ACL 2024 Findings & ICLR 2024 WS] An Evaluator VLM that is open-source, offers reproducible evaluation, and inexpensive to use. Specific…☆81Sep 13, 2024Updated last year
- Distilabel is a framework for synthetic data and AI feedback for engineers who need fast, reliable and scalable pipelines based on verifi…☆3,131Updated this week
- Repository for "Scaling Evaluation-time Compute with Reasoning Models as Process Evaluators"☆12Mar 25, 2025Updated 11 months ago
- Robust recipes to align language models with human and AI preferences☆5,527Sep 8, 2025Updated 6 months ago
- A framework for few-shot evaluation of language models.☆11,704Mar 5, 2026Updated 2 weeks ago
- An automatic evaluator for instruction-following language models. Human-validated, high-quality, cheap, and fast.☆1,961Aug 9, 2025Updated 7 months ago
- Arena-Hard-Auto: An automatic LLM benchmark.☆1,008Jun 21, 2025Updated 9 months ago
- [NeurIPS 2024] Train LLMs with diverse system messages reflecting individualized preferences to generalize to unseen system messages☆53Aug 10, 2025Updated 7 months ago
- AllenAI's post-training codebase☆3,629Updated this week
- Official repository for ORPO☆473May 31, 2024Updated last year
- The LLM Evaluation Framework☆14,115Mar 13, 2026Updated last week
- Tools for merging pretrained large language models.☆6,867Updated this week
- Go ahead and axolotl questions☆11,460Updated this week
- A library with extensible implementations of DPO, KTO, PPO, ORPO, and other human-aware loss functions (HALOs).☆908Sep 30, 2025Updated 5 months ago
- A lightweight library for generating synthetic instruction tuning datasets for your data without GPT.☆823Jul 15, 2025Updated 8 months ago
- Lighteval is your all-in-one toolkit for evaluating LLMs across multiple backends☆2,339Mar 9, 2026Updated last week
- The Universe of Evaluation. All about the evaluation for LLMs.☆233Jul 9, 2024Updated last year
- RewardBench: the first evaluation tool for reward models.☆704Feb 16, 2026Updated last month
- Stanford NLP Python library for Representation Finetuning (ReFT)☆1,564Mar 5, 2026Updated 2 weeks ago
- DSPy: The framework for programming—not prompting—language models☆32,853Updated this week
- Official repo for the paper PHUDGE: Phi-3 as Scalable Judge. Evaluate your LLMs with or without custom rubric, reference answer, absolute…☆52Jul 10, 2024Updated last year
- Easily use and train state of the art late-interaction retrieval methods (ColBERT) in any RAG pipeline. Designed for modularity and ease-…☆3,882May 17, 2025Updated 10 months ago
- Freeing data processing from scripting madness by providing a set of platform-agnostic customizable pipeline processing blocks.☆2,956Updated this week
- The official evaluation suite and dynamic data release for MixEval.☆255Nov 10, 2024Updated last year
- [NeurIPS 2024] SimPO: Simple Preference Optimization with a Reference-Free Reward☆948Feb 16, 2025Updated last year
- ☆11Sep 19, 2025Updated 6 months ago
- [EMNLP'23, ACL'24] To speed up LLMs' inference and enhance LLM's perceive of key information, compress the prompt and KV-Cache, which ach…☆5,937Oct 28, 2025Updated 4 months ago
- Supercharge Your LLM Application Evaluations 🚀☆13,008Feb 24, 2026Updated 3 weeks ago
- IntructIR, a novel benchmark specifically designed to evaluate the instruction following ability in information retrieval models. Our foc…☆32Jun 13, 2024Updated last year
- Benchmarking LLMs with Challenging Tasks from Real Users☆247Nov 3, 2024Updated last year
- Structured Outputs☆13,564Mar 9, 2026Updated last week
- A lightweight, low-dependency, unified API to use all common reranking and cross-encoder models.☆1,605Dec 20, 2025Updated 3 months ago
- Train transformer language models with reinforcement learning.☆17,697Updated this week
- PyTorch native post-training library☆5,707Updated this week
- [ICLR 2025 Spotlight] An open-sourced LLM judge for evaluating LLM-generated answers.☆423Feb 11, 2025Updated last year
- Gorilla: Training and Evaluating LLMs for Function Calls (Tool Calls)☆12,765Mar 11, 2026Updated last week
- Code for paper "G-Eval: NLG Evaluation using GPT-4 with Better Human Alignment"☆411Feb 4, 2024Updated 2 years ago