princeton-nlp / intercode
[NeurIPS 2023 D&B] Code repository for InterCode benchmark https://arxiv.org/abs/2306.14898
☆216Updated last year
Alternatives and similar repositories for intercode:
Users that are interested in intercode are comparing it to the libraries listed below
- Accepted by Transactions on Machine Learning Research (TMLR)☆126Updated 7 months ago
- CRUXEval: Code Reasoning, Understanding, and Execution Evaluation☆136Updated 6 months ago
- A set of utilities for running few-shot prompting experiments on large-language models☆120Updated last year
- ☆227Updated 8 months ago
- [NeurIPS 2022] 🛒WebShop: Towards Scalable Real-World Web Interaction with Grounded Language Agents☆336Updated 8 months ago
- ✨ RepoBench: Benchmarking Repository-Level Code Auto-Completion Systems - ICLR 2024☆162Updated 8 months ago
- [ICLR'24 Spotlight] A language model (LM)-based emulation framework for identifying the risks of LM agents with tool use☆142Updated last year
- 🌍 Repository for "AppWorld: A Controllable World of Apps and People for Benchmarking Interactive Coding Agent", ACL'24 Best Resource Pap…☆187Updated this week
- An Analytical Evaluation Board of Multi-turn LLM Agents [NeurIPS 2024 Oral]☆309Updated 11 months ago
- Code for the paper 🌳 Tree Search for Language Model Agents☆197Updated 9 months ago
- CrossCodeEval: A Diverse and Multilingual Benchmark for Cross-File Code Completion (NeurIPS 2023)☆139Updated 9 months ago
- ToolBench, an evaluation suite for LLM tool manipulation capabilities.☆150Updated last year
- Benchmarking LLMs with Challenging Tasks from Real Users☆221Updated 6 months ago
- Code for paper "LEVER: Learning to Verifiy Language-to-Code Generation with Execution" (ICML'23)☆86Updated last year
- A Collection of Competitive Text-Based Games for Language Model Evaluation and Reinforcement Learning☆150Updated this week
- ☆120Updated 7 months ago
- ☆115Updated 9 months ago
- Official Repo for ICLR 2024 paper MINT: Evaluating LLMs in Multi-turn Interaction with Tools and Language Feedback by Xingyao Wang*, Ziha…☆123Updated 11 months ago
- r2e: turn any github repository into a programming agent environment☆116Updated 2 weeks ago
- [ICML 2023] Data and code release for the paper "DS-1000: A Natural and Reliable Benchmark for Data Science Code Generation".☆242Updated 6 months ago
- Can Language Models Solve Olympiad Programming?☆116Updated 3 months ago
- Evaluating LLMs with fewer examples☆151Updated last year
- Code release for "Debating with More Persuasive LLMs Leads to More Truthful Answers"☆104Updated last year
- A hard gym for programming☆153Updated 9 months ago
- A banchmark list for evaluation of large language models.☆102Updated last week
- An extensible benchmark for evaluating large language models on planning☆355Updated last week
- ToolkenGPT: Augmenting Frozen Language Models with Massive Tools via Tool Embeddings - NeurIPS 2023 (oral)☆262Updated last year
- ☆287Updated 10 months ago
- VisualWebArena is a benchmark for multimodal agents.☆334Updated 5 months ago
- ☆121Updated 10 months ago