scicode-bench / SciCodeLinks
A benchmark that challenges language models to code solutions for scientific problems
☆143Updated last week
Alternatives and similar repositories for SciCode
Users that are interested in SciCode are comparing it to the libraries listed below
Sorting:
- Repository for the paper Stream of Search: Learning to Search in Language☆151Updated 8 months ago
- Evaluation of LLMs on latest math competitions☆171Updated 3 weeks ago
- ☆123Updated 7 months ago
- [ICML 2025] Flow of Reasoning: Training LLMs for Divergent Reasoning with Minimal Examples☆106Updated 2 months ago
- Can Language Models Solve Olympiad Programming?☆118Updated 8 months ago
- [ICLR'25] ScienceAgentBench: Toward Rigorous Assessment of Language Agents for Data-Driven Scientific Discovery☆103Updated last month
- Replicating O1 inference-time scaling laws☆90Updated 10 months ago
- ☆39Updated 6 months ago
- Archon provides a modular framework for combining different inference-time techniques and LMs with just a JSON config file.☆184Updated 7 months ago
- A simple unified framework for evaluating LLMs☆250Updated 5 months ago
- Code for NeurIPS'24 paper 'Grokked Transformers are Implicit Reasoners: A Mechanistic Journey to the Edge of Generalization'☆230Updated 2 months ago
- [COLM 2025] EvalTree: Profiling Language Model Weaknesses via Hierarchical Capability Trees☆24Updated 2 months ago
- Meta Agents Research Environments is a comprehensive platform designed to evaluate AI agents in dynamic, realistic scenarios. Unlike stat…☆282Updated 2 weeks ago
- ☆74Updated last month
- [NeurIPS 2023 D&B] Code repository for InterCode benchmark https://arxiv.org/abs/2306.14898☆225Updated last year
- Official repository for "Scaling Retrieval-Based Langauge Models with a Trillion-Token Datastore".☆216Updated 2 months ago
- RL Scaling and Test-Time Scaling (ICML'25)☆111Updated 8 months ago
- [COLM 2025] Official repository for R2E-Gym: Procedural Environment Generation and Hybrid Verifiers for Scaling Open-Weights SWE Agents☆166Updated 2 months ago
- The official repo for "TheoremQA: A Theorem-driven Question Answering dataset" (EMNLP 2023)☆34Updated last year
- ☆192Updated 5 months ago
- Functional Benchmarks and the Reasoning Gap☆89Updated last year
- Dynamic Cheatsheet: Test-Time Learning with Adaptive Memory☆74Updated 4 months ago
- RepoQA: Evaluating Long-Context Code Understanding☆117Updated 11 months ago
- Benchmarking LLMs with Challenging Tasks from Real Users☆241Updated 11 months ago
- ☆83Updated 8 months ago
- ☆115Updated 4 months ago
- Framework and toolkits for building and evaluating collaborative agents that can work together with humans.☆99Updated last week
- A banchmark list for evaluation of large language models.☆143Updated last month
- ☆100Updated last year
- CodeElo: Benchmarking Competition-level Code Generation of LLMs with Human-comparable Elo Ratings☆54Updated 8 months ago