RenzeLou / AAAR-1.0
The source code for running LLMs on the AAAR-1.0 benchmark.
☆11Updated this week
Related projects ⓘ
Alternatives and complementary repositories for AAAR-1.0
- MUFFIN: Curating Multi-Faceted Instructions for Improving Instruction-Following☆12Updated last week
- Codebase for Instruction Following without Instruction Tuning☆30Updated last month
- [ICLR'24 spotlight] Tool-Augmented Reward Modeling☆36Updated 8 months ago
- ☆15Updated 3 months ago
- [COLM'24] "How Easily do Irrelevant Inputs Skew the Responses of Large Language Models?"☆19Updated 3 weeks ago
- Improving Language Understanding from Screenshots. Paper: https://arxiv.org/abs/2402.14073☆26Updated 4 months ago
- The Good, The Bad, and The Greedy: Evaluation of LLMs Should Not Ignore Non-Determinism☆25Updated 3 months ago
- The official repository for the paper "From Zero to Hero: Examining the Power of Symbolic Tasks in Instruction Tuning".☆61Updated last year
- Source code for MMEvalPro, a more trustworthy and efficient benchmark for evaluating LMMs☆22Updated last month
- ☆29Updated this week
- GSM-Plus: Data, Code, and Evaluation for Enhancing Robust Mathematical Reasoning in Math Word Problems.☆46Updated 4 months ago
- Official implementation of the paper "From Complex to Simple: Enhancing Multi-Constraint Complex Instruction Following Ability of Large L…☆37Updated 4 months ago
- This the implementation of LeCo☆27Updated 3 months ago
- Official repository for paper "Weak-to-Strong Extrapolation Expedites Alignment"☆67Updated 5 months ago
- ☆19Updated 5 months ago
- InstructRAG: Instructing Retrieval-Augmented Generation via Self-Synthesized Rationales☆51Updated 3 weeks ago
- This is the official repository of the paper "OlympicArena: Benchmarking Multi-discipline Cognitive Reasoning for Superintelligent AI"☆85Updated last month
- This is the repo for our paper "Mr-Ben: A Comprehensive Meta-Reasoning Benchmark for Large Language Models"☆43Updated last week
- [ACL 2024 Findings] CriticBench: Benchmarking LLMs for Critique-Correct Reasoning☆20Updated 8 months ago
- Lightweight tool to identify Data Contamination in LLMs evaluation☆40Updated 8 months ago
- Source code of "Reasons to Reject? Aligning Language Models with Judgments"☆56Updated 8 months ago
- Repository for "Propagating Knowledge Updates to LMs Through Distillation" (NeurIPS 2023).☆24Updated 2 months ago
- Code implementation of synthetic continued pretraining☆54Updated last month
- [NeurIPS 2023] Repetition In Repetition Out: Towards Understanding Neural Text Degeneration from the Data Perspective☆29Updated last year
- The official repository of the Omni-MATH benchmark.☆47Updated last week
- Code for "Seeking Neural Nuggets: Knowledge Transfer in Large Language Models from a Parametric Perspective"☆29Updated 6 months ago
- Repository for NPHardEval, a quantified-dynamic benchmark of LLMs☆48Updated 7 months ago
- Evaluate the Quality of Critique☆35Updated 5 months ago
- Code and data for paper "Context-faithful Prompting for Large Language Models".☆39Updated last year
- ☆53Updated 2 months ago