felipemaiapolo / tinyBenchmarksLinks
Evaluating LLMs with fewer examples
☆158Updated last year
Alternatives and similar repositories for tinyBenchmarks
Users that are interested in tinyBenchmarks are comparing it to the libraries listed below
Sorting:
- Benchmarking LLMs with Challenging Tasks from Real Users☆226Updated 7 months ago
- ☆97Updated 11 months ago
- ☆180Updated 2 months ago
- Code accompanying "How I learned to start worrying about prompt formatting".☆105Updated 2 weeks ago
- Code and Data for "Long-context LLMs Struggle with Long In-context Learning" [TMLR2025]☆105Updated 4 months ago
- Co-LLM: Learning to Decode Collaboratively with Multiple Language Models☆115Updated last year
- Reproducible, flexible LLM evaluations☆213Updated last month
- ☆115Updated 4 months ago
- LongEmbed: Extending Embedding Models for Long Context Retrieval (EMNLP 2024)☆137Updated 7 months ago
- Official repository for "Scaling Retrieval-Based Langauge Models with a Trillion-Token Datastore".☆205Updated 2 weeks ago
- Stanford NLP Python library for benchmarking the utility of LLM interpretability methods☆95Updated 2 weeks ago
- Scalable Meta-Evaluation of LLMs as Evaluators☆42Updated last year
- ☆183Updated last year
- Positional Skip-wise Training for Efficient Context Window Extension of LLMs to Extremely Length (ICLR 2024)☆203Updated last year
- Official github repo for the paper "Compression Represents Intelligence Linearly" [COLM 2024]☆138Updated 9 months ago
- The official evaluation suite and dynamic data release for MixEval.☆242Updated 7 months ago
- open-source code for paper: Retrieval Head Mechanistically Explains Long-Context Factuality☆200Updated 10 months ago
- Functional Benchmarks and the Reasoning Gap☆87Updated 8 months ago
- The HELMET Benchmark☆154Updated 2 months ago
- ☆85Updated 7 months ago
- "Improving Mathematical Reasoning with Process Supervision" by OPENAI☆109Updated this week
- Code and example data for the paper: Rule Based Rewards for Language Model Safety☆188Updated 11 months ago
- Official code for "MAmmoTH2: Scaling Instructions from the Web" [NeurIPS 2024]☆144Updated 7 months ago
- OpenCoconut implements a latent reasoning paradigm where we generate thoughts before decoding.☆173Updated 5 months ago
- ☆65Updated last year
- Codebase accompanying the Summary of a Haystack paper.☆78Updated 9 months ago
- Code for the paper "Rethinking Benchmark and Contamination for Language Models with Rephrased Samples"☆303Updated last year
- Critique-out-Loud Reward Models☆66Updated 8 months ago
- A simple unified framework for evaluating LLMs☆219Updated 2 months ago
- ☆232Updated 10 months ago