nimasteryang / LingxiLinks
☆23Updated last month
Alternatives and similar repositories for Lingxi
Users that are interested in Lingxi are comparing it to the libraries listed below
Sorting:
- Agentless Lite: RAG-based SWE-Bench software engineering scaffold☆33Updated 2 months ago
- Harness used to benchmark aider against SWE Bench benchmarks☆72Updated last year
- ☆64Updated last month
- ☆101Updated last month
- Run SWE-bench evaluations remotely☆21Updated last month
- Official implementation of paper How to Understand Whole Repository? New SOTA on SWE-bench Lite (21.3%)☆86Updated 3 months ago
- Scaling Data for SWE-agents☆265Updated this week
- Enhanced fork of SWE-bench, tailored for OpenDevin's ecosystem.☆25Updated last year
- ☆97Updated 11 months ago
- ☆158Updated 10 months ago
- r2e: turn any github repository into a programming agent environment☆125Updated 2 months ago
- [ACL25' Findings] SWE-Dev is an SWE agent with a scalable test case construction pipeline.☆40Updated 2 weeks ago
- OrcaLoca: An LLM Agent Framework for Software Issue Localization [ICML 25]☆20Updated 2 months ago
- ☆42Updated 2 months ago
- Aider's refactoring benchmark exercises based on popular python repos☆74Updated 8 months ago
- Contains the model patches and the eval logs from the passing swe-bench-lite run.☆10Updated last year
- Official repository for R2E-Gym: Procedural Environment Generation and Hybrid Verifiers for Scaling Open-Weights SWE Agents☆77Updated 2 weeks ago
- Enhancing AI Software Engineering with Repository-level Code Graph☆185Updated 2 months ago
- Multi-SWE-bench: A Multilingual Benchmark for Issue Resolving☆195Updated this week
- Agent computer interface for AI software engineer.☆85Updated this week
- ✨ RepoBench: Benchmarking Repository-Level Code Auto-Completion Systems - ICLR 2024☆167Updated 10 months ago
- ☆41Updated 6 months ago
- Open sourced predictions, execution logs, trajectories, and results from model inference + evaluation runs on the SWE-bench task.☆186Updated this week
- Benchmark and research code for the paper SWEET-RL Training Multi-Turn LLM Agents onCollaborative Reasoning Tasks☆219Updated last month
- The evaluation benchmark on MCP servers☆134Updated last month
- Official Repo for The Paper "Talk Structurally, Act Hierarchically: A Collaborative Framework for LLM Multi-Agent Systems"☆54Updated 4 months ago
- ☆211Updated last month
- Framework and toolkits for building and evaluating collaborative agents that can work together with humans.☆86Updated 2 months ago
- 🌍 Repository for "AppWorld: A Controllable World of Apps and People for Benchmarking Interactive Coding Agent", ACL'24 Best Resource Pap…☆215Updated last month
- Reasoning by Communicating with Agents☆29Updated last month