qtli / GSM-Plus
GSM-Plus: Data, Code, and Evaluation for Enhancing Robust Mathematical Reasoning in Math Word Problems.
☆55Updated 6 months ago
Alternatives and similar repositories for GSM-Plus:
Users that are interested in GSM-Plus are comparing it to the libraries listed below
- ☆26Updated 3 weeks ago
- [AAAI 2025 oral] Evaluating Mathematical Reasoning Beyond Accuracy☆44Updated last month
- Evaluate the Quality of Critique☆35Updated 7 months ago
- [EMNLP 2024] Source code for the paper "Learning Planning-based Reasoning with Trajectory Collection and Process Rewards Synthesizing".☆62Updated this week
- BeHonest: Benchmarking Honesty in Large Language Models☆30Updated 5 months ago
- Official repository for paper "Weak-to-Strong Extrapolation Expedites Alignment"☆71Updated 7 months ago
- 🍼 Official implementation of Dynamic Data Mixing Maximizes Instruction Tuning for Mixture-of-Experts☆35Updated 3 months ago
- We introduce ScaleQuest, a scalable, novel and cost-effective data synthesis method to unleash the reasoning capability of LLMs.☆58Updated 2 months ago
- Resources for our ACL 2023 paper: Distilling Script Knowledge from Large Language Models for Constrained Language Planning☆35Updated last year
- Analyzing LLM Alignment via Token distribution shift☆14Updated 11 months ago
- [ICLR'24 spotlight] Tool-Augmented Reward Modeling☆44Updated 3 weeks ago
- ☆71Updated 7 months ago
- Towards Systematic Measurement for Long Text Quality☆31Updated 4 months ago
- ☆47Updated 9 months ago
- Grade-School Math with Irrelevant Context (GSM-IC) benchmark is an arithmetic reasoning dataset built upon GSM8K, by adding irrelevant se…☆58Updated last year
- ☆93Updated last year
- ☆27Updated 10 months ago
- The Good, The Bad, and The Greedy: Evaluation of LLMs Should Not Ignore Non-Determinism☆26Updated 6 months ago
- Code and data for paper "Context-faithful Prompting for Large Language Models".☆39Updated last year
- [ACL 2024 Findings] CriticBench: Benchmarking LLMs for Critique-Correct Reasoning☆20Updated 10 months ago
- Code for M4LE: A Multi-Ability Multi-Range Multi-Task Multi-Domain Long-Context Evaluation Benchmark for Large Language Models☆22Updated 5 months ago
- GPT as Human☆18Updated last month
- ☆56Updated 4 months ago
- The official repository of the Omni-MATH benchmark.☆66Updated 3 weeks ago
- Code & Data for our Paper "Alleviating Hallucinations of Large Language Models through Induced Hallucinations"☆62Updated 10 months ago
- AbstainQA, ACL 2024☆25Updated 3 months ago
- Benchmarking Benchmark Leakage in Large Language Models☆47Updated 7 months ago
- The repository of the project "Fine-tuning Large Language Models with Sequential Instructions", code base comes from open-instruct and LA…☆29Updated last month
- ☆61Updated 9 months ago
- [ICML'24] TroVE: Inducing Verifiable and Efficient Toolboxes for Solving Programmatic Tasks☆23Updated 3 months ago