SakanaAI / ALE-BenchLinks
The official repository of ALE-Bench
☆156Updated this week
Alternatives and similar repositories for ALE-Bench
Users that are interested in ALE-Bench are comparing it to the libraries listed below
Sorting:
- An AI benchmark for creative, human-like problem solving using Sudoku variants☆158Updated last month
- [ICLR 2026] Official PyTorch Implementation of RLP: Reinforcement as a Pretraining Objective☆231Updated 2 weeks ago
- ☆37Updated 8 months ago
- Training teachers with reinforcement learning able to make LLMs learn how to reason for test time scaling.☆358Updated 7 months ago
- This repo contains the source code for the paper "Evolution Strategies at Scale: LLM Fine-Tuning Beyond Reinforcement Learning"☆292Updated 2 months ago
- Archon provides a modular framework for combining different inference-time techniques and LMs with just a JSON config file.☆189Updated 11 months ago
- ☆394Updated last week
- The Automated LLM Speedrunning Benchmark measures how well LLM agents can reproduce previous innovations and discover new ones in languag…☆128Updated 4 months ago
- ☆90Updated 3 months ago
- AIRA-dojo: a framework for developing and evaluating AI research agents☆125Updated 2 weeks ago
- ☆106Updated 7 months ago
- ☆229Updated 11 months ago
- ☆117Updated last year
- Can Language Models Solve Olympiad Programming?☆123Updated last year
- AlgoTune is a NeurIPS 2025 benchmark made up of 154 math, physics, and computer science problems. The goal is write code that solves each…☆88Updated last week
- accompanying material for sleep-time compute paper☆119Updated 9 months ago
- [ACL 2024] Do Large Language Models Latently Perform Multi-Hop Reasoning?☆90Updated 10 months ago
- [ICLR 2026] RPG: KL-Regularized Policy Gradient (https://arxiv.org/abs/2505.17508)☆65Updated 2 weeks ago
- [ICML 2025] Flow of Reasoning: Training LLMs for Divergent Reasoning with Minimal Examples☆120Updated last week
- ☆148Updated this week
- [NeurIPS 2025 Spotlight] Co-Evolving LLM Coder and Unit Tester via Reinforcement Learning☆149Updated 4 months ago
- Systematic evaluation framework that automatically rates overthinking behavior in large language models.☆96Updated 8 months ago
- ☆123Updated 11 months ago
- [ICLR 2026] Learning to Reason without External Rewards☆389Updated 2 weeks ago
- [COLM 2025] Code for Paper: Learning Adaptive Parallel Reasoning with Language Models☆139Updated last month
- ☆88Updated 7 months ago
- EvaByte: Efficient Byte-level Language Models at Scale☆115Updated 9 months ago
- Reinforcing General Reasoning without Verifiers☆96Updated 7 months ago
- Open source interpretability artefacts for R1.☆170Updated 9 months ago
- Implementation of SOAR☆49Updated 4 months ago