openai / evalsView on GitHub
Evals is a framework for evaluating LLMs and LLM systems, and an open-source registry of benchmarks.
18,129Mar 26, 2026Updated last week

Alternatives and similar repositories for evals

Users that are interested in evals are comparing it to the libraries listed below. We may earn a commission when you buy through links labeled 'Ad' on this page.

Sorting:

Are these results useful?