A tool for evaluating LLMs
☆428Mar 15, 2026Updated last month
Alternatives and similar repositories for bench
Users that are interested in bench are comparing it to the libraries listed below. We may earn a commission when you buy through links labeled 'Ad' on this page.
Sorting:
- Python SDK for running evaluations on LLM generated responses☆299Jun 6, 2025Updated 10 months ago
- Supercharge Your LLM Application Evaluations 🚀☆13,415Feb 24, 2026Updated last month
- Fiddler Auditor is a tool to evaluate language models.☆189Mar 11, 2024Updated 2 years ago
- The LLM Evaluation Framework☆14,728Apr 9, 2026Updated last week
- Continuous Integration for LLM powered applications☆257Aug 11, 2023Updated 2 years ago
- Managed hosting for WordPress and PHP on Cloudways • AdManaged hosting for WordPress, Magento, Laravel, or PHP apps, on multiple cloud providers. Deploy in minutes on Cloudways by DigitalOcean.
- AI Observability & Evaluation☆9,284Updated this week
- An open-source visual programming environment for battle-testing prompts to LLMs.☆2,971Apr 6, 2026Updated last week
- React Hooks for Keyless AI☆25Apr 13, 2023Updated 3 years ago
- Open-source tools for prompt testing and experimentation, with support for both LLMs (e.g. OpenAI, LLaMA) and vector databases (e.g. Chro…☆3,032Feb 11, 2026Updated 2 months ago
- Adding guardrails to large language models.☆6,675Apr 3, 2026Updated 2 weeks ago
- Hosted embedding platform to discover, evaluate, and retrieve embeddings☆73Sep 21, 2023Updated 2 years ago
- Distilabel is a framework for synthetic data and AI feedback for engineers who need fast, reliable and scalable pipelines based on verifi…☆3,178Updated this week
- NeMo Guardrails is an open-source toolkit for easily adding programmable guardrails to LLM-based conversational systems.☆5,986Updated this week
- Automated Evaluation of RAG Systems☆702Mar 28, 2025Updated last year
- Managed hosting for WordPress and PHP on Cloudways • AdManaged hosting for WordPress, Magento, Laravel, or PHP apps, on multiple cloud providers. Deploy in minutes on Cloudways by DigitalOcean.
- DSPy: The framework for programming—not prompting—language models☆33,649Apr 13, 2026Updated last week
- [ICLR 2025 Spotlight] An open-sourced LLM judge for evaluating LLM-generated answers.☆426Feb 11, 2025Updated last year
- Go ahead and axolotl questions☆11,688Updated this week
- Retrieval Augmented Generation (RAG) chatbot powered by Weaviate☆7,646Jul 14, 2025Updated 9 months ago
- Bayesian Optimization as a Coverage Tool for Evaluating LLMs. Accurate evaluation (benchmarking) that's 10 times faster with just a few l…☆289Mar 18, 2026Updated last month
- A framework for few-shot evaluation of language models.☆12,138Apr 8, 2026Updated last week
- Evaluation and Tracking for LLM Experiments and AI Agents☆3,240Apr 9, 2026Updated last week
- A guidance language for controlling large language models.☆21,381Apr 10, 2026Updated last week
- Promptimize is a prompt engineering evaluation and testing toolkit.☆493Mar 16, 2026Updated last month
- GPUs on demand by Runpod - Special Offer Available • AdRun AI, ML, and HPC workloads on powerful cloud GPUs—without limits or wasted spend. Deploy GPUs in under a minute and pay by the second.
- Evaluate your LLM's response with Prometheus and GPT4 💯☆1,066Apr 25, 2025Updated 11 months ago
- Python client library for improving your LLM app accuracy☆96Feb 11, 2025Updated last year
- The papers are organized according to our survey: Evaluating Large Language Models: A Comprehensive Survey.☆799May 8, 2024Updated last year
- Sample notebooks and prompts for LLM evaluation☆160Nov 2, 2025Updated 5 months ago
- Evals is a framework for evaluating LLMs and LLM systems, and an open-source registry of benchmarks.☆18,227Updated this week
- LLM Prompt Injection Detector☆1,459Aug 7, 2024Updated last year
- 🐢 Open-Source Evaluation & Testing library for LLM Agents☆5,273Updated this week
- structured outputs for llms☆12,749Updated this week
- Structured Outputs☆13,657Mar 26, 2026Updated 3 weeks ago
- Bare Metal GPUs on DigitalOcean Gradient AI • AdPurpose-built for serious AI teams training foundational models, running large-scale inference, and pushing the boundaries of what's possible.
- Holistic Evaluation of Language Models (HELM) is an open source Python framework created by the Center for Research on Foundation Models …☆2,741Apr 10, 2026Updated last week
- Gorilla: Training and Evaluating LLMs for Function Calls (Tool Calls)☆12,814Mar 23, 2026Updated 3 weeks ago
- A python command-line tool to download & manage MLX AI models from Hugging Face.☆19Aug 26, 2024Updated last year
- VectorFlow is a high volume vector embedding pipeline that ingests raw data, transforms it into vectors and writes it to a vector DB of y…☆701May 16, 2024Updated last year
- 🪢 Open source LLM engineering platform: LLM Observability, metrics, evals, prompt management, playground, datasets. Integrates with Open…☆25,055Updated this week
- Data-Driven Evaluation for LLM-Powered Applications☆516Jan 22, 2025Updated last year
- An efficient, to-the-point, and easy-to-use checklist to following when deploying an ML model into production.☆30Jan 25, 2023Updated 3 years ago