THUDM / SWE-DevLinks
[ACL25' Findings] SWE-Dev is an SWE agent with a scalable test case construction pipeline.
☆56Updated 4 months ago
Alternatives and similar repositories for SWE-Dev
Users that are interested in SWE-Dev are comparing it to the libraries listed below
Sorting:
- ☆125Updated 6 months ago
- ☆123Updated 5 months ago
- SWE-Swiss: A Multi-Task Fine-Tuning and RL Recipe for High-Performance Issue Resolution☆97Updated 2 months ago
- WideSearch: Benchmarking Agentic Broad Info-Seeking☆102Updated last month
- ☆54Updated last year
- [COLM 2025] Official repository for R2E-Gym: Procedural Environment Generation and Hybrid Verifiers for Scaling Open-Weights SWE Agents☆197Updated 4 months ago
- Archon provides a modular framework for combining different inference-time techniques and LMs with just a JSON config file.☆189Updated 8 months ago
- [ACL 2025] Agentic Reward Modeling: Integrating Human Preferences with Verifiable Correctness Signals for Reliable Reward Systems☆112Updated 5 months ago
- RL Scaling and Test-Time Scaling (ICML'25)☆112Updated 10 months ago
- This is the official implementation for paper "PENCIL: Long Thoughts with Short Memory".☆68Updated 6 months ago
- [ICML 2025] Flow of Reasoning: Training LLMs for Divergent Reasoning with Minimal Examples☆112Updated 4 months ago
- RepoQA: Evaluating Long-Context Code Understanding☆124Updated last year
- ☆80Updated 3 weeks ago
- ☆105Updated 11 months ago
- [NeurIPS'25 D&B] Mind2Web-2 Benchmark: Evaluating Agentic Search with Agent-as-a-Judge☆89Updated 3 weeks ago
- [NeurIPS 2025 Spotlight] ReasonFlux-Coder: Open-Source LLM Coders with Co-Evolving Reinforcement Learning☆133Updated 2 months ago
- ☆77Updated 8 months ago
- CodeElo: Benchmarking Competition-level Code Generation of LLMs with Human-comparable Elo Ratings☆56Updated 9 months ago
- A Comprehensive Benchmark for Software Development.☆119Updated last year
- StepCoder: Improve Code Generation with Reinforcement Learning from Compiler Feedback☆73Updated last year
- NaturalCodeBench (Findings of ACL 2024)☆67Updated last year
- Systematic evaluation framework that automatically rates overthinking behavior in large language models.☆94Updated 6 months ago
- The code for the paper ROUTERBENCH: A Benchmark for Multi-LLM Routing System☆150Updated last year
- SSRL: Self-Search Reinforcement Learning☆156Updated 3 months ago
- Run SWE-bench evaluations remotely☆44Updated 3 months ago
- Code for the paper: CodeTree: Agent-guided Tree Search for Code Generation with Large Language Models☆29Updated 7 months ago
- Moatless Testbeds allows you to create isolated testbed environments in a Kubernetes cluster where you can apply code changes through git…☆14Updated 7 months ago
- Code for paper "Optima: Optimizing Effectiveness and Efficiency for LLM-Based Multi-Agent System"☆66Updated last year
- ☆46Updated 5 months ago
- e☆41Updated 7 months ago