ssu-humane / HerOLinks
The code for HerO: a fact-checking pipeline based on open LLMs (the runner-up in AVeriTeC)
☆11Updated 7 months ago
Alternatives and similar repositories for HerO
Users that are interested in HerO are comparing it to the libraries listed below
Sorting:
- Source Code of Paper "GPTScore: Evaluate as You Desire"☆257Updated 2 years ago
- A package to evaluate factuality of long-form generation. Original implementation of our EMNLP 2023 paper "FActScore: Fine-grained Atomic…☆394Updated 6 months ago
- Code for paper "G-Eval: NLG Evaluation using GPT-4 with Better Human Alignment"☆385Updated last year
- ☆68Updated 11 months ago
- ☆85Updated 10 months ago
- Tools for checking ACL paper submissions☆836Updated last month
- Multilingual Large Language Models Evaluation Benchmark☆132Updated last year
- Awesome LLM for NLG Evaluation Papers☆25Updated last year
- The implementation of "RQUGE: Reference-Free Metric for Evaluating Question Generation by Answering the Question" [ACL 2023]☆16Updated last year
- BARTScore: Evaluating Generated Text as Text Generation☆358Updated 3 years ago
- ACL 2022: An Empirical Survey of the Effectiveness of Debiasing Techniques for Pre-trained Language Models.☆149Updated 2 months ago
- SemEval2024-task8: Multidomain, Multimodel and Multilingual Machine-Generated Text Detection☆77Updated last year
- Code and data for Marked Personas (ACL 2023)☆28Updated 2 years ago
- Repository for EMNLP 2022 Paper: Towards a Unified Multi-Dimensional Evaluator for Text Generation☆211Updated last year
- ☆189Updated 3 months ago
- This is a repository for sharing papers in the field of persona-based conversational AI. The related source code for each paper is linked…☆167Updated last year
- Codebase, data and models for the SummaC paper in TACL☆102Updated 8 months ago
- Github repository for "RAGTruth: A Hallucination Corpus for Developing Trustworthy Retrieval-Augmented Language Models"☆205Updated 10 months ago
- Unofficial re-implementation of "Trusting Your Evidence: Hallucinate Less with Context-aware Decoding"☆31Updated 11 months ago
- This is the repository of HaluEval, a large-scale hallucination evaluation benchmark for Large Language Models.☆517Updated last year
- ConvGQR: Generative Query Reformulation for Conversational Search. A codebase for ACL 2023 accepted paper.☆32Updated last year
- This repository contains the data and code introduced in the paper "CrowS-Pairs: A Challenge Dataset for Measuring Social Biases in Maske…☆125Updated last year
- Data and info for the paper "ParaDetox: Text Detoxification with Parallel Data"☆32Updated 6 months ago
- Simple replication of DPR (Dense Passage Retrieval)☆48Updated last year
- ☆154Updated 2 years ago
- Code and data for Koo et al's ACL 2024 paper "Benchmarking Cognitive Biases in Large Language Models as Evaluators"☆21Updated last year
- Official code for the ACL 2024 paper: Chat Vector: A Simple Approach to Equip LLMs with Instruction Following and Model Alignment in New …☆55Updated last year
- ACL 2023: Evaluating Open-Domain Question Answering in the Era of Large Language Models☆47Updated last year
- [EMNLP 2023] Enabling Large Language Models to Generate Text with Citations. Paper: https://arxiv.org/abs/2305.14627☆497Updated last year
- BLEnD: A Benchmark for LLMs on Everyday Knowledge in Diverse Cultures and Languages☆37Updated 2 months ago