Aider-AI / refactor-benchmark
Aider's refactoring benchmark exercises based on popular python repos
☆70Updated 6 months ago
Alternatives and similar repositories for refactor-benchmark:
Users that are interested in refactor-benchmark are comparing it to the libraries listed below
- Harness used to benchmark aider against SWE Bench benchmarks☆71Updated 10 months ago
- Agent computer interface for AI software engineer.☆68Updated this week
- A DSPy-based implementation of the tree of thoughts method (Yao et al., 2023) for generating persuasive arguments☆78Updated 7 months ago
- Just a bunch of benchmark logs for different LLMs☆119Updated 9 months ago
- Official homepage for "Self-Harmonized Chain of Thought" (NAACL 2025)☆90Updated 3 months ago
- proof-of-concept of Cursor's Instant Apply feature☆78Updated 8 months ago
- Coding problems used in aider's polyglot benchmark☆110Updated 4 months ago
- Sandboxed code execution for AI agents, locally or on the cloud. Massively parallel, easy to extend. Powering SWE-agent and more.☆170Updated this week
- an implementation of Self-Extend, to expand the context window via grouped attention☆119Updated last year
- ☆73Updated last year
- Leveraging DSPy for AI-driven task understanding and solution generation, the Self-Discover Framework automates problem-solving through r…☆60Updated 9 months ago
- A better way of testing, inspecting, and analyzing AI Agent traces.☆35Updated last week
- Synthetic data derived by templating, few shot prompting, transformations on public domain corpora, and monte carlo tree search.☆32Updated 2 months ago
- Client Code Examples, Use Cases and Benchmarks for Enterprise h2oGPTe RAG-Based GenAI Platform☆87Updated last week
- ☆48Updated last year
- ☆38Updated last year
- ReDel is a toolkit for researchers and developers to build, iterate on, and analyze recursive multi-agent systems. (EMNLP 2024 Demo)☆76Updated last month
- ☆155Updated 8 months ago
- ☆22Updated 10 months ago
- auto fine tune of models with synthetic data☆75Updated last year
- Mixing Language Models with Self-Verification and Meta-Verification☆104Updated 4 months ago
- A framework for evaluating function calls made by LLMs☆37Updated 9 months ago
- Score LLM pretraining data with classifiers☆55Updated last year
- ☆20Updated last year
- Chat Markup Language conversation library☆55Updated last year
- Simple Graph Memory for AI applications☆84Updated 9 months ago
- Track the progress of LLM context utilisation☆54Updated 3 weeks ago
- Optimizing Causal LMs through GRPO with weighted reward functions and automated hyperparameter tuning using Optuna☆39Updated 3 months ago
- Using various instructor clients evaluating the quality and capabilities of extractions and reasoning.☆51Updated 7 months ago
- Open sourced predictions, execution logs, trajectories, and results from model inference + evaluation runs on the SWE-bench task.☆169Updated last month