ZhonghaoJiang / CoSILLinks
[ASE 2025] CoSIL: Software Issue Localization via LLM-Driven Code Repository Graph Searching
β18Updated last month
Alternatives and similar repositories for CoSIL
Users that are interested in CoSIL are comparing it to the libraries listed below
Sorting:
- Must-read papers on Repository-level Code Generation & Issue Resolution π₯β251Updated last month
- Enhancing AI Software Engineering with Repository-level Code Graphβ250Updated 10 months ago
- [NeurIPS 2025 D&B Spotlight] Scaling Data for SWE-agentsβ551Updated this week
- Multi-SWE-bench: A Multilingual Benchmark for Issue Resolvingβ317Updated last month
- Official implementation of paper How to Understand Whole Repository? New SOTA on SWE-bench Lite (21.3%)β95Updated 10 months ago
- β132Updated 8 months ago
- [ACL25' Findings] SWE-Dev is an SWE agent with a scalable test case construction pipeline.β58Updated 6 months ago
- CodeRAG-Bench: Can Retrieval Augment Code Generation?β167Updated last year
- ResearcherBench: Evaluating Deep AI Research Systems on the Frontiers of Scientific Inquiryβ45Updated last month
- β46Updated 3 months ago
- β331Updated 6 months ago
- SWE-Swiss: A Multi-Task Fine-Tuning and RL Recipe for High-Performance Issue Resolutionβ104Updated 4 months ago
- Language Models for Code Completion: a Practical Evaluationβ13Updated 2 years ago
- β58Updated last year
- A curated list of papers related to constrained decoding of LLM, along with their relevant code and resources.β320Updated 3 weeks ago
- [ICLR 2025] TidalDecode: A Fast and Accurate LLM Decoding with Position Persistent Sparse Attentionβ52Updated 6 months ago
- CrossCodeEval: A Diverse and Multilingual Benchmark for Cross-File Code Completion (NeurIPS 2023)β171Updated 5 months ago
- A comprehensive code domain benchmark review of LLM researches.β195Updated 4 months ago
- [COLM 2025] Official repository for R2E-Gym: Procedural Environment Generation and Hybrid Verifiers for Scaling Open-Weights SWE Agentsβ236Updated 7 months ago
- Repo-Level Code generation papersβ233Updated last month
- The repository for paper "DebugBench: "Evaluating Debugging Capability of Large Language Models".β85Updated last year
- A Comprehensive Benchmark for Software Development.β127Updated last year
- RepoQA: Evaluating Long-Context Code Understandingβ128Updated last year
- Live-SWE-agent: live, runtime self-evolving software engineering agentβ240Updated 3 weeks ago
- β32Updated 3 months ago
- Meta Agents Research Environments is a comprehensive platform designed to evaluate AI agents in dynamic, realistic scenarios. Unlike statβ¦β427Updated 2 weeks ago
- Official repository for our paper "FullStack Bench: Evaluating LLMs as Full Stack Coders"β112Updated 9 months ago
- (ACL 2025 Main) Code for MultiAgentBench : Evaluating the Collaboration and Competition of LLM agents https://www.arxiv.org/pdf/2503.019β¦β217Updated 3 months ago
- Official Repo of CudaForgeβ59Updated 2 months ago
- [NeurIPS '25] GSO: Challenging Software Optimization Tasks for Evaluating SWE-Agentsβ63Updated last week