openai / simple-evalsLinks
☆4,340Updated 6 months ago
Alternatives and similar repositories for simple-evals
Users that are interested in simple-evals are comparing it to the libraries listed below
Sorting:
- AllenAI's post-training codebase☆3,562Updated this week
- Doing simple retrieval from LLM models at various context lengths to measure accuracy☆2,167Updated last year
- SWE-bench: Can Language Models Resolve Real-world Github Issues?☆4,232Updated this week
- Distilabel is a framework for synthetic data and AI feedback for engineers who need fast, reliable and scalable pipelines based on verifi…☆3,074Updated 2 weeks ago
- Lighteval is your all-in-one toolkit for evaluating LLMs across multiple backends☆2,291Updated 2 weeks ago
- An automatic evaluator for instruction-following language models. Human-validated, high-quality, cheap, and fast.☆1,940Updated 6 months ago
- DataComp for Language Models☆1,413Updated 5 months ago
- PyTorch native post-training library☆5,660Updated last week
- Democratizing Reinforcement Learning for LLMs☆5,081Updated this week
- Tools for merging pretrained large language models.☆6,761Updated 2 weeks ago
- A library for advanced large language model reasoning☆2,328Updated 7 months ago
- TextGrad: Automatic ''Differentiation'' via Text -- using large language models to backpropagate textual gradients. Published in Nature.☆3,341Updated 6 months ago
- Arena-Hard-Auto: An automatic LLM benchmark.☆994Updated 7 months ago
- Holistic Evaluation of Language Models (HELM) is an open source Python framework created by the Center for Research on Foundation Models …☆2,662Updated this week
- A unified evaluation framework for large language models☆2,773Updated 2 weeks ago
- MLE-bench is a benchmark for measuring how well AI agents perform at machine learning engineering☆1,301Updated 3 weeks ago
- Measuring Massive Multitask Language Understanding | ICLR 2021☆1,550Updated 2 years ago
- An Open Large Reasoning Model for Real-World Solutions☆1,533Updated this week
- [NeurIPS 2024] OSWorld: Benchmarking Multimodal Agents for Open-Ended Tasks in Real Computer Environments☆2,528Updated last week
- Modeling, training, eval, and inference code for OLMo☆6,305Updated 2 months ago
- ☆4,112Updated last year
- A framework for few-shot evaluation of language models.☆11,358Updated this week
- A Comprehensive Benchmark to Evaluate LLMs as Agents (ICLR'24)☆3,151Updated 2 months ago
- Our library for RL environments + evals☆3,809Updated this week
- Code and Data for Tau-Bench☆1,087Updated 5 months ago
- LiveBench: A Challenging, Contamination-Free LLM Benchmark☆1,032Updated this week
- Curated list of datasets and tools for post-training.☆4,229Updated 2 months ago
- Fully open data curation for reasoning models☆2,206Updated 2 months ago
- A collection of benchmarks and datasets for evaluating LLM.☆550Updated last year
- Awesome Reasoning LLM Tutorial/Survey/Guide☆2,286Updated 3 months ago