IAAR-Shanghai / UHGEvalLinks
[ACL 2024] User-friendly evaluation framework: Eval Suite & Benchmarks: UHGEval, HaluEval, HalluQA, etc.
☆167Updated 6 months ago
Alternatives and similar repositories for UHGEval
Users that are interested in UHGEval are comparing it to the libraries listed below
Sorting:
- Controllable Text Generation for Large Language Models: A Survey☆175Updated 9 months ago
- This includes the original implementation of CtrlA: Adaptive Retrieval-Augmented Generation via Inherent Control.☆60Updated 7 months ago
- Explore concepts like Self-Correct, Self-Refine, Self-Improve, Self-Contradict, Self-Play, and Self-Knowledge, alongside o1-like reasonin…☆168Updated 5 months ago
- notes for Multi-hop Reading Comprehension and open-domain question answering☆86Updated 3 years ago
- Grimoire is All You Need for Enhancing Large Language Models☆115Updated last year
- ☆130Updated 2 months ago
- We leverage 14 datasets as OOD test data and conduct evaluations on 8 NLU tasks over 21 popularly used models. Our findings confirm that …☆93Updated last year
- [EMNLP 2023] FreeAL: Towards Human-Free Active Learning in the Era of Large Language Models☆89Updated last year
- The framework to prune LLMs to any size and any config.☆92Updated last year
- MPLSandbox is an out-of-the-box multi-programming language sandbox designed to provide unified and comprehensive feedback from compiler a…☆176Updated last month
- Codebase for Iterative DPO Using Rule-based Rewards☆245Updated last month
- SQL-o1: A Self-Reward Heuristic Dynamic Search Method for Text-to-SQL☆106Updated last week
- ✨ A synthetic dataset generation framework that produces diverse coding questions and verifiable solutions - all in one framwork☆226Updated 2 months ago
- [ACL 2024 Main] NewsBench: A Systematic Evaluation Framework for Assessing Editorial Capabilities of Large Language Models in Chinese Jou…☆31Updated 11 months ago
- [ACL2024 Findings] Towards Better Question Generation in QA-based Event Extraction