lmarena / copilot-arenaLinks
☆340Updated last week
Alternatives and similar repositories for copilot-arena
Users that are interested in copilot-arena are comparing it to the libraries listed below
Sorting:
- A system that tries to resolve all issues on a github repo with OpenHands.☆117Updated last year
- Coding problems used in aider's polyglot benchmark☆194Updated 11 months ago
- Finetune Llama-3-8b on the MathInstruct dataset☆114Updated last year
- Verify Precision of all Kimi K2 API Vendor☆461Updated 3 weeks ago
- Together Open Deep Research☆355Updated 7 months ago
- Building open version of OpenAI o1 via reasoning traces (Groq, ollama, Anthropic, Gemini, OpenAI, Azure supported) Demo: https://hugging…☆187Updated last year
- Letting Claude Code develop his own MCP tools :)☆123Updated 9 months ago
- Open-source resources on agents for computer use.☆385Updated 2 months ago
- Agent computer interface for AI software engineer.☆114Updated 2 months ago
- ☆434Updated last year
- Contains the prompts we use to talk to various LLMs for different utilities inside the editor☆83Updated last year
- Sandboxed code execution for AI agents, locally or on the cloud. Massively parallel, easy to extend. Powering SWE-agent and more.☆388Updated this week
- GRadient-INformed MoE☆264Updated last year
- ☆180Updated 11 months ago
- proof-of-concept of Cursor's Instant Apply feature☆87Updated last year
- Instantly calculate the maximum size of quantized language models that can fit in your available RAM, helping you optimize your models fo…☆242Updated 7 months ago
- ☆79Updated 2 months ago
- Testing baseline LLMs performance across various models☆325Updated last week
- Open Agent Computer Interface☆89Updated last year
- A Text-Based Environment for Interactive Debugging☆282Updated last week
- Code to accompany the Universal Deep Research paper (https://arxiv.org/abs/2509.00244)☆450Updated 3 months ago
- Harness used to benchmark aider against SWE Bench benchmarks☆78Updated last year
- [NeurIPS 2025 D&B Spotlight] Scaling Data for SWE-agents☆479Updated last week
- ☆67Updated 6 months ago
- DevQualityEval: An evaluation benchmark 📈 and framework to compare and evolve the quality of code generation of LLMs.☆182Updated 6 months ago
- ☆59Updated 10 months ago
- ☆191Updated last year
- Super basic implementation (gist-like) of RLMs with REPL environments.☆278Updated last month
- Claude Deep Research config for Claude Code.☆222Updated 8 months ago
- A comprehensive set of LLM benchmark scores and provider prices. (deprecated, read more in README)☆354Updated last month