GAIR-NLP / MetaCritiqueLinks
Evaluate the Quality of Critique
☆35Updated last year
Alternatives and similar repositories for MetaCritique
Users that are interested in MetaCritique are comparing it to the libraries listed below
Sorting:
- GSM-Plus: Data, Code, and Evaluation for Enhancing Robust Mathematical Reasoning in Math Word Problems.☆62Updated 11 months ago
- AbstainQA, ACL 2024☆26Updated 8 months ago
- Search, Verify and Feedback: Towards Next Generation Post-training Paradigm of Foundation Models via Verifier Engineering☆59Updated 6 months ago
- BeHonest: Benchmarking Honesty in Large Language Models☆34Updated 10 months ago
- ☆30Updated 5 months ago
- ☆59Updated 9 months ago
- [ACL'24] Code and data of paper "When is Tree Search Useful for LLM Planning? It Depends on the Discriminator"☆54Updated last year
- Revisiting Mid-training in the Era of RL Scaling☆56Updated last month
- [ICLR'24 spotlight] Tool-Augmented Reward Modeling☆50Updated 2 weeks ago
- [NAACL 2025] The official implementation of paper "Learning From Failure: Integrating Negative Examples when Fine-tuning Large Language M…☆26Updated last year
- RAG-RewardBench: Benchmarking Reward Models in Retrieval Augmented Generation for Preference Alignment☆16Updated 6 months ago
- [AAAI 2025 oral] Evaluating Mathematical Reasoning Beyond Accuracy☆63Updated 6 months ago
- ☆41Updated last year
- ☆44Updated 10 months ago
- ☆13Updated 11 months ago
- Code and data for paper "Context-faithful Prompting for Large Language Models".☆40Updated 2 years ago
- [ACL 2024] Code for "MoPS: Modular Story Premise Synthesis for Open-Ended Automatic Story Generation"☆38Updated 11 months ago
- The official repository of "Improving Large Language Models via Fine-grained Reinforcement Learning with Minimum Editing Constraint"☆38Updated last year
- Scalable Meta-Evaluation of LLMs as Evaluators☆42Updated last year
- LongHeads: Multi-Head Attention is Secretly a Long Context Processor☆29Updated last year
- The code of “Improving Weak-to-Strong Generalization with Scalable Oversight and Ensemble Learning”☆16Updated last year
- This is a unified platform for implementing and evaluating test-time reasoning mechanisms in Large Language Models (LLMs).☆18Updated 5 months ago
- The repository of the project "Fine-tuning Large Language Models with Sequential Instructions", code base comes from open-instruct and LA…☆29Updated 6 months ago
- [ACL 2025] Are Your LLMs Capable of Stable Reasoning?☆25Updated 3 months ago
- The rule-based evaluation subset and code implementation of Omni-MATH☆22Updated 6 months ago
- ☆46Updated 7 months ago
- Towards Systematic Measurement for Long Text Quality☆35Updated 9 months ago
- Benchmarking Benchmark Leakage in Large Language Models☆51Updated last year
- Watch Every Step! LLM Agent Learning via Iterative Step-level Process Refinement (EMNLP 2024 Main Conference)☆57Updated 8 months ago
- ☆14Updated last year