bigcode-project / selfcodealignLinks
[NeurIPS'24] SelfCodeAlign: Self-Alignment for Code Generation
☆309Updated 5 months ago
Alternatives and similar repositories for selfcodealign
Users that are interested in selfcodealign are comparing it to the libraries listed below
Sorting:
- Code for the paper "Rethinking Benchmark and Contamination for Language Models with Rephrased Samples"☆306Updated last year
- RepoQA: Evaluating Long-Context Code Understanding☆113Updated 9 months ago
- Scaling Data for SWE-agents☆328Updated this week
- CRUXEval: Code Reasoning, Understanding, and Execution Evaluation☆151Updated 9 months ago
- Run evaluation on LLMs using human-eval benchmark☆417Updated last year
- 🐙 OctoPack: Instruction Tuning Code Large Language Models☆472Updated 5 months ago
- A simple unified framework for evaluating LLMs☆229Updated 3 months ago
- Experiments on speculative sampling with Llama models☆128Updated 2 years ago
- ☆108Updated 2 months ago
- ✨ RepoBench: Benchmarking Repository-Level Code Auto-Completion Systems - ICLR 2024☆168Updated 11 months ago
- Open sourced predictions, execution logs, trajectories, and results from model inference + evaluation runs on the SWE-bench task.☆197Updated 3 weeks ago
- OpenCoconut implements a latent reasoning paradigm where we generate thoughts before decoding.☆173Updated 6 months ago
- ☆311Updated last year
- EvolKit is an innovative framework designed to automatically enhance the complexity of instructions used for fine-tuning Large Language M…☆229Updated 9 months ago
- Archon provides a modular framework for combining different inference-time techniques and LMs with just a JSON config file.☆174Updated 4 months ago
- Code for Paper: Training Software Engineering Agents and Verifiers with SWE-Gym [ICML 2025]☆513Updated this week
- The official evaluation suite and dynamic data release for MixEval.☆242Updated 8 months ago
- ☆270Updated 2 years ago
- Benchmarking LLMs with Challenging Tasks from Real Users☆233Updated 9 months ago
- [NeurIPS 2023 D&B] Code repository for InterCode benchmark https://arxiv.org/abs/2306.14898☆223Updated last year
- ☆159Updated 11 months ago
- Official repository for the paper "LiveCodeBench: Holistic and Contamination Free Evaluation of Large Language Models for Code"☆608Updated 2 weeks ago
- Fine-tune SantaCoder for Code/Text Generation.☆192Updated 2 years ago
- Reproducible, flexible LLM evaluations☆226Updated 3 weeks ago
- The code and data for "MMLU-Pro: A More Robust and Challenging Multi-Task Language Understanding Benchmark" [NeurIPS 2024]☆264Updated 5 months ago
- ☆84Updated 2 years ago
- BABILong is a benchmark for LLM evaluation using the needle-in-a-haystack approach.☆208Updated 2 months ago
- Spherical Merge Pytorch/HF format Language Models with minimal feature loss.☆135Updated last year
- Official codebase for "SWE-RL: Advancing LLM Reasoning via Reinforcement Learning on Open Software Evolution"☆571Updated 4 months ago
- The official repo for "LLoCo: Learning Long Contexts Offline"☆118Updated last year