benchflow-ai / benchflowLinks
AI benchmark runtime framework that allows you to integrate and evaluate AI tasks using Docker-based benchmarks.
☆169Updated 3 weeks ago
Alternatives and similar repositories for benchflow
Users that are interested in benchflow are comparing it to the libraries listed below
Sorting:
- The LLM abstraction layer for modern AI agent applications.☆499Updated last week
- Agent computer interface for AI software engineer.☆114Updated last month
- Challenges for general-purpose web-browsing AI agents☆67Updated 7 months ago
- ⚖️ Awesome LLM Judges ⚖️☆148Updated 8 months ago
- Commit0: Library Generation from Scratch☆174Updated 8 months ago
- Sandboxed code execution for AI agents, locally or on the cloud. Massively parallel, easy to extend. Powering SWE-agent and more.☆402Updated last week
- [NeurIPS 2025 D&B Spotlight] Scaling Data for SWE-agents☆509Updated this week
- ☆175Updated 10 months ago
- AWM: Agent Workflow Memory☆376Updated 2 weeks ago
- Curated collection of community environments☆200Updated this week
- [ACL 2024] Do Large Language Models Latently Perform Multi-Hop Reasoning?☆85Updated 9 months ago
- Harbor is a framework for running agent evaluations and creating and using RL environments.☆306Updated this week
- Beating the GAIA benchmark with Transformers Agents. 🚀☆144Updated 10 months ago
- A clean, modular SDK for building AI agents with OpenHands V1.☆391Updated this week
- Computer Agent Arena: Test & compare AI agents in real desktop apps & web environments. Code/data coming soon!☆51Updated 9 months ago
- Meta Agents Research Environments is a comprehensive platform designed to evaluate AI agents in dynamic, realistic scenarios. Unlike stat…☆411Updated last month
- SWE-Bench Pro: Can AI Agents Solve Long-Horizon Software Engineering Tasks?☆240Updated this week
- Prompt design in Python☆63Updated last year
- Prompt-to-Leaderboard☆270Updated 8 months ago
- GRPO training code which scales to 32xH100s for long horizon terminal/coding tasks. Base agent is now the top Qwen3 agent on Stanford's T…☆323Updated 4 months ago
- ☆63Updated 6 months ago
- Coding problems used in aider's polyglot benchmark☆199Updated last year
- ☆128Updated 7 months ago
- τ²-Bench: Evaluating Conversational Agents in a Dual-Control Environment☆598Updated 3 weeks ago
- [NAACL2025] LiteWebAgent: The Open-Source Suite for VLM-Based Web-Agent Applications☆141Updated 6 months ago
- Framework and toolkits for building and evaluating collaborative agents that can work together with humans.☆117Updated last month
- SkillsBench evaluates how well skills work and how effective agents are at using them☆166Updated this week
- 🦀️ CRAB: Cross-environment Agent Benchmark for Multimodal Language Model Agents. https://crab.camel-ai.org/☆388Updated last week
- ☆68Updated 7 months ago
- ☆130Updated 8 months ago