EuniAI / awesome-code-agentsLinks
A curated list of products, benchmarks, and research papers on autonomous code agents. Beyond coding — they're redefining how software changes the world.
☆55Updated this week
Alternatives and similar repositories for awesome-code-agents
Users that are interested in awesome-code-agents are comparing it to the libraries listed below
Sorting:
- The repository for paper "DebugBench: "Evaluating Debugging Capability of Large Language Models".☆85Updated last year
- ☆25Updated 4 months ago
- SWE-Swiss: A Multi-Task Fine-Tuning and RL Recipe for High-Performance Issue Resolution☆101Updated 3 months ago
- Reinforcement Learning for Repository-Level Code Completion☆43Updated last year
- R-Judge: Benchmarking Safety Risk Awareness for LLM Agents (EMNLP Findings 2024)☆93Updated 7 months ago
- ☆33Updated 3 months ago
- A Comprehensive Benchmark for Software Development.☆124Updated last year
- Reproducing R1 for Code with Reliable Rewards☆278Updated 7 months ago
- An Evolving Code Generation Benchmark Aligned with Real-world Code Repositories☆67Updated last year
- [NeurIPS 2025 D&B] 🚀 SWE-bench Goes Live!☆146Updated last week
- ☆51Updated 9 months ago
- ☆83Updated 4 months ago
- A comprehensive code domain benchmark review of LLM researches.☆176Updated 3 months ago
- Must-read papers on Repository-level Code Generation & Issue Resolution 🔥☆227Updated this week
- Official implementation of paper How to Understand Whole Repository? New SOTA on SWE-bench Lite (21.3%)☆95Updated 9 months ago
- Neural Code Intelligence Survey 2024-25; Reading lists and resources☆279Updated 5 months ago
- Repo-Level Code generation papers☆226Updated last week
- Official github repo for AutoDetect, an automated weakness detection framework for LLMs.☆44Updated last year
- CodeRAG-Bench: Can Retrieval Augment Code Generation?☆163Updated last year
- ☆32Updated 6 months ago
- Code for the AAAI 2023 paper "CodeAttack: Code-based Adversarial Attacks for Pre-Trained Programming Language Models☆33Updated 2 years ago
- A novel approach to improve the safety of large language models, enabling them to transition effectively from unsafe to safe state.☆73Updated 7 months ago
- ☆23Updated last year
- Repoformer: Selective Retrieval for Repository-Level Code Completion (ICML 2024)☆64Updated 6 months ago
- 【ACL 2024】 SALAD benchmark & MD-Judge☆169Updated 9 months ago
- ☆15Updated last year
- NaturalCodeBench (Findings of ACL 2024)☆68Updated last year
- Official repository for our paper "FullStack Bench: Evaluating LLMs as Full Stack Coders"☆107Updated 7 months ago
- Code for ACL (main) paper "JumpCoder: Go Beyond Autoregressive Coder via Online Modification"☆27Updated last year
- Code for paper "Defending aginast LLM Jailbreaking via Backtranslation"☆33Updated last year