ConsequentAI / fnevalLinks
Functional Benchmarks and the Reasoning Gap
☆86Updated 8 months ago
Alternatives and similar repositories for fneval
Users that are interested in fneval are comparing it to the libraries listed below
Sorting:
- Implementation of the paper: "AssistantBench: Can Web Agents Solve Realistic and Time-Consuming Tasks?"☆56Updated 5 months ago
- Anchored Preference Optimization and Contrastive Revisions: Addressing Underspecification in Alignment☆57Updated 9 months ago
- ☆114Updated 3 months ago
- ☆58Updated 2 weeks ago
- Replicating O1 inference-time scaling laws☆87Updated 6 months ago
- CausalGym: Benchmarking causal interpretability methods on linguistic tasks☆42Updated 6 months ago
- ☆34Updated 2 months ago
- ☆27Updated 3 weeks ago
- Repository for the paper Stream of Search: Learning to Search in Language☆146Updated 3 months ago
- Archon provides a modular framework for combining different inference-time techniques and LMs with just a JSON config file.☆173Updated 2 months ago
- Official repo for Learning to Reason for Long-Form Story Generation☆58Updated last month
- ☆49Updated 6 months ago
- ☆60Updated last year
- Evaluating LLMs with fewer examples☆156Updated last year
- Evaluating LLMs with CommonGen-Lite☆90Updated last year
- ☆75Updated last month
- ☆120Updated 8 months ago
- Code for the ICLR 2024 paper "How to catch an AI liar: Lie detection in black-box LLMs by asking unrelated questions"☆70Updated 11 months ago
- The GitHub repo for Goal Driven Discovery of Distributional Differences via Language Descriptions☆70Updated 2 years ago
- Just a bunch of benchmark logs for different LLMs☆118Updated 10 months ago
- SWE Arena☆33Updated last month
- ☆130Updated 2 months ago
- Public Inflection Benchmarks☆68Updated last year
- Code and data for the paper "Why think step by step? Reasoning emerges from the locality of experience"☆60Updated last month
- OpenCoconut implements a latent reasoning paradigm where we generate thoughts before decoding.☆171Updated 4 months ago
- Scalable Meta-Evaluation of LLMs as Evaluators☆42Updated last year
- Simple replication of [ColBERT-v1](https://arxiv.org/abs/2004.12832).☆80Updated last year
- EvaByte: Efficient Byte-level Language Models at Scale☆98Updated last month
- Repository for NPHardEval, a quantified-dynamic benchmark of LLMs☆54Updated last year
- Official repository for "Scaling Retrieval-Based Langauge Models with a Trillion-Token Datastore".☆201Updated 3 weeks ago