bigscience-workshop / evaluation
Code and Data for Evaluation WG
☆41Updated 2 years ago
Alternatives and similar repositories for evaluation:
Users that are interested in evaluation are comparing it to the libraries listed below
- A Dataset for Tuning and Evaluation of Sentence Simplification Models with Multiple Rewriting Transformations☆54Updated 2 years ago
- ☆74Updated 3 years ago
- Implementation of Marge, Pre-training via Paraphrasing, in Pytorch☆75Updated 4 years ago
- Faithfulness and factuality annotations of XSum summaries from our paper "On Faithfulness and Factuality in Abstractive Summarization" (h…☆81Updated 4 years ago
- ☆22Updated 3 years ago
- ☆46Updated 5 years ago
- QED: A Framework and Dataset for Explanations in Question Answering☆115Updated 3 years ago
- Codebase, data and models for the Keep it Simple paper at ACL2021☆38Updated last year
- Statistics on multilingual datasets☆17Updated 2 years ago
- REALSumm: Re-evaluating Evaluation in Text Summarization☆71Updated 2 years ago
- This is a repository for the paper on testing inductive bias with scaled-down RoBERTa models.☆20Updated 3 years ago
- This is the official repository for NAACL 2021, "XOR QA: Cross-lingual Open-Retrieval Question Answering".☆79Updated 3 years ago
- Research code for the paper "How Good is Your Tokenizer? On the Monolingual Performance of Multilingual Language Models"☆26Updated 3 years ago
- EMNLP 2021 Tutorial: Multi-Domain Multilingual Question Answering☆38Updated 3 years ago
- This repository accompanies our paper “Do Prompt-Based Models Really Understand the Meaning of Their Prompts?”☆85Updated 2 years ago
- Automatic metrics for GEM tasks☆63Updated 2 years ago
- SacreROUGE is a library dedicated to the use and development of text generation evaluation metrics with an emphasis on summarization.☆140Updated 2 years ago
- ☆58Updated last year
- Code to support the paper "Question and Answer Test-Train Overlap in Open-Domain Question Answering Datasets"☆66Updated 3 years ago
- Few-shot NLP benchmark for unified, rigorous eval☆91Updated 2 years ago
- ☆37Updated 3 years ago
- ☆20Updated 2 years ago
- ☆38Updated 4 years ago
- Repro is a library for easily running code from published papers via Docker.☆40Updated last year
- An original implementation of EMNLP 2020, "AmbigQA: Answering Ambiguous Open-domain Questions"☆117Updated 2 years ago
- ☆67Updated 3 years ago
- codebase for the Text-based NP Enrichment (TNE) paper☆19Updated 10 months ago
- Code and dataset for the EMNLP 2021 Finding paper "Can NLI Models Verify QA Systems’ Predictions?"☆25Updated last year
- Code to reproduce the experiments from the paper.☆101Updated last year
- A benchmark for understanding and evaluating rationales: http://www.eraserbenchmark.com/☆96Updated 2 years ago