d223302 / LLM-EvaluationLinks
Can Large Language Models Be an Alternative to Human Evaluations?
☆9Updated last year
Alternatives and similar repositories for LLM-Evaluation
Users that are interested in LLM-Evaluation are comparing it to the libraries listed below
Sorting:
- This repository contains the dataset and code for "WiCE: Real-World Entailment for Claims in Wikipedia" in EMNLP 2023.☆41Updated last year
- ☆44Updated last year
- Easy-to-use framework for evaluating cross-lingual consistency of factual knowledge (Supported LLaMA, BLOOM, mT5, RoBERTa, etc.) Paper he…☆23Updated 3 months ago
- This repository accompanies our paper “Do Prompt-Based Models Really Understand the Meaning of Their Prompts?”☆85Updated 3 years ago
- ☆48Updated 2 years ago
- ☆82Updated 2 years ago
- ☆48Updated 2 years ago
- The geometry of multilingual language model representations (EMNLP 2022).☆21Updated 2 years ago
- ☆26Updated 2 years ago
- ☆27Updated last year
- Codebase, data and models for the SummaC paper in TACL☆96Updated 4 months ago
- FRANK: Factuality Evaluation Benchmark☆56Updated 2 years ago
- Dataset, metrics, and models for TACL 2023 paper MACSUM: Controllable Summarization with Mixed Attributes.☆34Updated last year
- ☆43Updated 2 years ago
- ☆15Updated 2 years ago
- Easy-to-use MIRAGE code for faithful answer attribution in RAG applications. Paper: https://aclanthology.org/2024.emnlp-main.347/☆24Updated 3 months ago
- Detect hallucinated tokens for conditional sequence generation.☆64Updated 3 years ago
- Github repository for "FELM: Benchmarking Factuality Evaluation of Large Language Models" (NeurIPS 2023)☆59Updated last year
- NAACL 2024: SeaEval for Multilingual Foundation Models: From Cross-Lingual Alignment to Cultural Reasoning☆25Updated 3 months ago
- ☆39Updated 2 years ago
- ☆58Updated 3 years ago
- Faithfulness and factuality annotations of XSum summaries from our paper "On Faithfulness and Factuality in Abstractive Summarization" (h…☆82Updated 4 years ago
- Data for evaluating gender bias in coreference resolution systems.☆77Updated 6 years ago
- ☆75Updated last year
- Dataset associated with "BOLD: Dataset and Metrics for Measuring Biases in Open-Ended Language Generation" paper☆80Updated 4 years ago
- ☆21Updated 6 months ago
- ☆22Updated last year
- ☆100Updated 2 years ago
- ☆17Updated last week
- WikiWhy is a new benchmark for evaluating LLMs' ability to explain between cause-effect relationships. It is a QA dataset containing 9000…☆47Updated last year