OpenAutoCoder / AgentlessLinks
Agentlessπ±: an agentless approach to automatically solve software development problems
β1,986Updated last year
Alternatives and similar repositories for Agentless
Users that are interested in Agentless are comparing it to the libraries listed below
Sorting:
- β617Updated 3 months ago
- LDB: A Large Language Model Debugger via Verifying Runtime Execution Step by Step (ACL'24)β572Updated last year
- Official Repo for ICML 2024 paper "Executable Code Actions Elicit Better LLM Agents" by Xingyao Wang, Yangyi Chen, Lifan Yuan, Yizhe Zhanβ¦β1,502Updated last year
- [ICLR 2025] Automated Design of Agentic Systemsβ1,477Updated 10 months ago
- AgentCoder: multi-agent code generation framework.β368Updated last month
- AIDE: AI-Driven Exploration in the Space of Code. The machine Learning engineering agent that automates AI R&D.β1,099Updated last month
- Code and Data for Tau-Benchβ1,033Updated 4 months ago
- End-to-end Generative Optimization for AI Agentsβ699Updated 2 weeks ago
- Agent driven automation starting with the web. Try it: https://www.emergence.ai/web-automation-apiβ1,195Updated 6 months ago
- A project structure aware autonomous software engineer aiming for autonomous program improvement. Resolved 37.3% tasks (pass@1) in SWE-beβ¦β3,044Updated 8 months ago
- OO for LLMsβ883Updated last week
- An LLM-powered repository agent designed to assist developers and teams in generating documentation and understanding repositories quicklβ¦β863Updated last year
- An agent benchmark with tasks in a simulated software company.β611Updated last month
- [NeurIPS'25] Official codebase for "SWE-RL: Advancing LLM Reasoning via Reinforcement Learning on Open Software Evolution"β637Updated 9 months ago
- An self-improving embodied conversational agent seamlessly integrated into the operating system to automate our daily tasks.β1,713Updated last year
- π©ββοΈ Agent-as-a-Judge: The Magic for Open-Endednessβ693Updated 7 months ago
- Rigourous evaluation of LLM-synthesized code - NeurIPS 2023 & COLM 2024β1,661Updated 2 months ago
- MLE-bench is a benchmark for measuring how well AI agents perform at machine learning engineeringβ1,236Updated last week
- Streamlines and simplifies prompt design for both developers and non-technical users with a low code approach.β1,129Updated 2 months ago
- SWE-bench: Can Language Models Resolve Real-world Github Issues?β4,017Updated last week
- The #1 open-source SWE-bench Verified implementationβ843Updated 6 months ago
- Code for "WebVoyager: WebVoyager: Building an End-to-End Web Agent with Large Multimodal Models"β985Updated last year
- Optimizing inference proxy for LLMsβ3,250Updated this week
- ππͺ BrowserGym, a Gym environment for web task automationβ1,046Updated last week
- β434Updated last year
- Sandboxed code execution for AI agents, locally or on the cloud. Massively parallel, easy to extend. Powering SWE-agent and more.β396Updated this week
- This repo contains the dataset and code for the paper "SWE-Lancer: Can Frontier LLMs Earn $1 Million from Real-World Freelance Software Eβ¦β1,438Updated 5 months ago
- Windows Agent Arena (WAA) πͺ is a scalable OS platform for testing and benchmarking of multi-modal AI agents.β802Updated 7 months ago
- Code for Paper: Training Software Engineering Agents and Verifiers with SWE-Gym [ICML 2025]β607Updated 4 months ago
- [NeurIPS 2024] OSWorld: Benchmarking Multimodal Agents for Open-Ended Tasks in Real Computer Environmentsβ2,411Updated this week