google-deepmind / questbenchLinks
☆37Updated 8 months ago
Alternatives and similar repositories for questbench
Users that are interested in questbench are comparing it to the libraries listed below
Sorting:
- [ICML 2025] Flow of Reasoning: Training LLMs for Divergent Reasoning with Minimal Examples☆120Updated last week
- ☆123Updated 11 months ago
- ☆35Updated 8 months ago
- Repository for the paper Stream of Search: Learning to Search in Language☆153Updated last year
- Official Repo for InSTA: Towards Internet-Scale Training For Agents☆56Updated 6 months ago
- ☆90Updated 3 months ago
- accompanying material for sleep-time compute paper☆119Updated 9 months ago
- [ACL 2025] Agentic Reward Modeling: Integrating Human Preferences with Verifiable Correctness Signals for Reliable Reward Systems☆125Updated 7 months ago
- A testbed for agents and environments that can automatically improve models through data generation.☆28Updated 11 months ago
- Systematic evaluation framework that automatically rates overthinking behavior in large language models.☆96Updated 8 months ago
- Learning to Retrieve by Trying - Source code for Grounding by Trying: LLMs with Reinforcement Learning-Enhanced Retrieval☆51Updated last year
- ☆228Updated 11 months ago
- Implementation of the paper: "AssistantBench: Can Web Agents Solve Realistic and Time-Consuming Tasks?"☆69Updated last year
- [ICLR 2026] RPG: KL-Regularized Policy Gradient (https://arxiv.org/abs/2505.17508)☆65Updated last week
- ☆99Updated last year
- Reinforcing General Reasoning without Verifiers☆96Updated 7 months ago
- ☆112Updated last year
- ☆33Updated last year
- Archon provides a modular framework for combining different inference-time techniques and LMs with just a JSON config file.☆189Updated 11 months ago
- official implementation of paper "Process Reward Model with Q-value Rankings"☆65Updated last year
- The official implementation of Self-Exploring Language Models (SELM)☆63Updated last year
- Process Reward Models That Think☆78Updated 2 months ago
- Replicating O1 inference-time scaling laws☆92Updated last year
- Anchored Preference Optimization and Contrastive Revisions: Addressing Underspecification in Alignment☆61Updated last year
- Verifiers for LLM Reinforcement Learning☆80Updated 9 months ago
- ☆141Updated 4 months ago
- Can Language Models Solve Olympiad Programming?☆123Updated last year
- Learning from preferences is a common paradigm for fine-tuning language models. Yet, many algorithmic design decisions come into play. Ou…☆32Updated last year
- WONDERBREAD benchmark + dataset for BPM tasks☆34Updated 6 months ago
- ☆49Updated 10 months ago