google-deepmind / alphaevolve_resultsLinks
☆246Updated 5 months ago
Alternatives and similar repositories for alphaevolve_results
Users that are interested in alphaevolve_results are comparing it to the libraries listed below
Sorting:
- Evaluation of LLMs on latest math competitions☆200Updated last month
- ShinkaEvolve: Towards Open-Ended and Sample-Efficient Program Evolution☆730Updated last week
- ☆569Updated 6 months ago
- ☆99Updated 2 months ago
- ☆478Updated 4 months ago
- Training teachers with reinforcement learning able to make LLMs learn how to reason for test time scaling.☆353Updated 5 months ago
- Open source interpretability artefacts for R1.☆164Updated 7 months ago
- Library for text-to-text regression, applicable to any input string representation and allows pretraining and fine-tuning over multiple r…☆298Updated last week
- ☆218Updated 8 months ago
- RLP: Reinforcement as a Pretraining Objective☆210Updated 2 months ago
- Research code artifacts for Code World Model (CWM) including inference tools, reproducibility, and documentation.☆756Updated 2 months ago
- Public repository for "The Surprising Effectiveness of Test-Time Training for Abstract Reasoning"☆340Updated last month
- Repository for Zochi's Research☆294Updated 3 weeks ago
- ☆320Updated 2 months ago
- Code for NeurIPS'24 paper 'Grokked Transformers are Implicit Reasoners: A Mechanistic Journey to the Edge of Generalization'☆234Updated 4 months ago
- ☆161Updated 3 months ago
- OpenAI Frontier Evals☆957Updated last week
- Code for the paper: "Learning to Reason without External Rewards"☆382Updated 5 months ago
- A framework to study AI models in Reasoning, Alignment, and use of Memory (RAM).☆331Updated this week
- ☆163Updated 3 weeks ago
- Testing baseline LLMs performance across various models☆330Updated last week
- ☆201Updated 3 months ago
- [NeurIPS 2025 D&B Spotlight] Scaling Data for SWE-agents☆485Updated this week
- MLGym A New Framework and Benchmark for Advancing AI Research Agents☆576Updated 4 months ago
- Meta Agents Research Environments is a comprehensive platform designed to evaluate AI agents in dynamic, realistic scenarios. Unlike stat…☆391Updated 3 weeks ago
- Open-source release accompanying Gao et al. 2025☆121Updated 2 weeks ago
- accompanying material for sleep-time compute paper☆118Updated 7 months ago
- Technical report of Kimina-Prover Preview.☆347Updated 5 months ago
- ☆275Updated 7 months ago
- A Collection of Competitive Text-Based Games for Language Model Evaluation and Reinforcement Learning☆322Updated last month