KbsdJames / omni-math-rule
The rule-based evaluation subset and code implementation of Omni-MATH
☆17Updated 3 months ago
Alternatives and similar repositories for omni-math-rule:
Users that are interested in omni-math-rule are comparing it to the libraries listed below
- The official repository of the Omni-MATH benchmark.☆77Updated 3 months ago
- RAG-RewardBench: Benchmarking Reward Models in Retrieval Augmented Generation for Preference Alignment☆16Updated 3 months ago
- ☆29Updated 2 months ago
- ☆59Updated 6 months ago
- BeHonest: Benchmarking Honesty in Large Language Models☆31Updated 7 months ago
- [ICLR 25 Oral] RM-Bench: Benchmarking Reward Models of Language Models with Subtlety and Style☆27Updated this week
- This is the repo for our paper "Mr-Ben: A Comprehensive Meta-Reasoning Benchmark for Large Language Models"☆47Updated 4 months ago
- [AAAI 2025 oral] Evaluating Mathematical Reasoning Beyond Accuracy☆55Updated 3 months ago
- GSM-Plus: Data, Code, and Evaluation for Enhancing Robust Mathematical Reasoning in Math Word Problems.☆55Updated 8 months ago
- Code for Paper: Teaching Language Models to Critique via Reinforcement Learning☆84Updated last month
- ☆13Updated 8 months ago
- Towards Systematic Measurement for Long Text Quality☆33Updated 6 months ago
- ☆43Updated 4 months ago
- Suri: Multi-constraint instruction following for long-form text generation (EMNLP’24)☆22Updated 4 months ago
- We introduce ScaleQuest, a scalable, novel and cost-effective data synthesis method to unleash the reasoning capability of LLMs.☆60Updated 4 months ago
- [ICLR'24 spotlight] Tool-Augmented Reward Modeling☆45Updated 2 months ago
- The implementation of paper "LLM Critics Help Catch Bugs in Mathematics: Towards a Better Mathematical Verifier with Natural Language Fee…☆38Updated 8 months ago
- 🍼 Official implementation of Dynamic Data Mixing Maximizes Instruction Tuning for Mixture-of-Experts☆38Updated 5 months ago
- Evaluate the Quality of Critique☆35Updated 9 months ago
- The Good, The Bad, and The Greedy: Evaluation of LLMs Should Not Ignore Non-Determinism☆28Updated 8 months ago
- The code and data for the paper JiuZhang3.0☆42Updated 9 months ago
- [NAACL 2025] Source code for MMEvalPro, a more trustworthy and efficient benchmark for evaluating LMMs☆23Updated 5 months ago
- [ACL 2024 Findings] CriticBench: Benchmarking LLMs for Critique-Correct Reasoning☆24Updated last year
- Official implementation of the paper "From Complex to Simple: Enhancing Multi-Constraint Complex Instruction Following Ability of Large L…☆47Updated 9 months ago
- Official repository for paper "Weak-to-Strong Extrapolation Expedites Alignment"☆73Updated 9 months ago
- Visual and Embodied Concepts evaluation benchmark☆21Updated last year
- LongHeads: Multi-Head Attention is Secretly a Long Context Processor☆29Updated 11 months ago
- Official completion of “Training on the Benchmark Is Not All You Need”.☆30Updated 2 months ago
- Search, Verify and Feedback: Towards Next Generation Post-training Paradigm of Foundation Models via Verifier Engineering☆58Updated 3 months ago
- Code for ICLR 2025 Paper "What is Wrong with Perplexity for Long-context Language Modeling?"☆44Updated last month