arcprize / model_baseline
Testing baseline LLMs performance across various models
☆260Updated last week
Alternatives and similar repositories for model_baseline
Users that are interested in model_baseline are comparing it to the libraries listed below
Sorting:
- Open source interpretability artefacts for R1.☆109Updated 3 weeks ago
- ☆247Updated last month
- A Collection of Competitive Text-Based Games for Language Model Evaluation and Reinforcement Learning☆156Updated this week
- A comprehensive repository of reasoning tasks for LLMs (and beyond)☆439Updated 7 months ago
- ☆111Updated 4 months ago
- procedural reasoning datasets☆580Updated this week
- OpenCoconut implements a latent reasoning paradigm where we generate thoughts before decoding.☆172Updated 4 months ago
- ☆125Updated last month
- Aidan Bench attempts to measure <big_model_smell> in LLMs.☆299Updated 3 weeks ago
- Build your own visual reasoning model☆362Updated this week
- Atropos is a Language Model Reinforcement Learning Environments framework for collecting and evaluating LLM trajectories through diverse …☆357Updated this week
- Code for Paper: Training Software Engineering Agents and Verifiers with SWE-Gym [ICML 2025]☆455Updated last week
- Official codebase for "SWE-RL: Advancing LLM Reasoning via Reinforcement Learning on Open Software Evolution"☆517Updated 2 months ago
- prime-rl is a codebase for decentralized RL training at scale☆211Updated this week
- Public repository for "The Surprising Effectiveness of Test-Time Training for Abstract Reasoning"☆307Updated 5 months ago
- Code for NeurIPS'24 paper 'Grokked Transformers are Implicit Reasoners: A Mechanistic Journey to the Edge of Generalization'☆191Updated 5 months ago
- ☆97Updated 7 months ago
- ☆150Updated 2 months ago
- Archon provides a modular framework for combining different inference-time techniques and LMs with just a JSON config file.☆173Updated 2 months ago
- SkyRL-v0: Train Real-World Long-Horizon Agents via Reinforcement Learning☆261Updated this week
- MLGym A New Framework and Benchmark for Advancing AI Research Agents☆492Updated this week
- A simple unified framework for evaluating LLMs☆211Updated last month
- Verifiers for LLM Reinforcement Learning☆953Updated this week
- ⚖️ Awesome LLM Judges ⚖️☆97Updated 2 weeks ago
- Long context evaluation for large language models☆208Updated 2 months ago
- Memory layers use a trainable key-value lookup mechanism to add extra parameters to a model without increasing FLOPs. Conceptually, spars…☆327Updated 5 months ago
- A benchmark that challenges language models to code solutions for scientific problems☆119Updated this week
- Exploring Applications of GRPO☆212Updated last week
- Draw more samples☆189Updated 10 months ago
- ☆148Updated 2 months ago