amazon-science / SWE-PolyBenchLinks
SWE-PolyBench: A multi-language benchmark for repository level evaluation of coding agents
☆77Updated this week
Alternatives and similar repositories for SWE-PolyBench
Users that are interested in SWE-PolyBench are comparing it to the libraries listed below
Sorting:
- Run SWE-bench evaluations remotely☆51Updated 5 months ago
- Open sourced predictions, execution logs, trajectories, and results from model inference + evaluation runs on the SWE-bench task.☆246Updated last week
- Sandboxed code execution for AI agents, locally or on the cloud. Massively parallel, easy to extend. Powering SWE-agent and more.☆430Updated this week
- [NeurIPS 2025 D&B Spotlight] Scaling Data for SWE-agents☆538Updated this week
- ☆223Updated this week
- Harbor is a framework for running agent evaluations and creating and using RL environments.☆542Updated this week
- A Text-Based Environment for Interactive Debugging☆293Updated last week
- Harness used to benchmark aider against SWE Bench benchmarks☆79Updated last year
- ☆132Updated 8 months ago
- A clean, modular SDK for building AI agents with OpenHands V1.☆476Updated last week
- Archon provides a modular framework for combining different inference-time techniques and LMs with just a JSON config file.☆189Updated 11 months ago
- Inference-time scaling for LLMs-as-a-judge.☆328Updated 3 months ago
- ☆237Updated 2 months ago
- Public repository containing METR's DVC pipeline for eval data analysis☆199Updated last week
- ☆59Updated last year
- Agent computer interface for AI software engineer.☆116Updated 2 months ago
- ☆106Updated last year
- CodeSage: Code Representation Learning At Scale (ICLR 2024)☆116Updated last year
- The Granite Guardian models are designed to detect risks in prompts and responses.☆130Updated 4 months ago
- [ICLR'25] BigCodeBench: Benchmarking Code Generation Towards AGI☆475Updated last month
- SWE-Bench Pro: Can AI Agents Solve Long-Horizon Software Engineering Tasks?☆259Updated last month
- accompanying material for sleep-time compute paper☆119Updated 9 months ago
- Coding problems used in aider's polyglot benchmark☆199Updated last year
- Data recipes and robust infrastructure for training AI agents☆94Updated this week
- Tutorial for building LLM router☆244Updated last year
- A benchmark for LLMs on complicated tasks in the terminal☆1,494Updated 2 weeks ago
- [NeurIPS '25] GSO: Challenging Software Optimization Tasks for Evaluating SWE-Agents☆63Updated this week
- τ²-Bench: Evaluating Conversational Agents in a Dual-Control Environment☆717Updated last week
- ☆236Updated 3 months ago
- ☆137Updated 10 months ago