neuro-symbolic-ai / reasoning_with_nle_emnlp_2024
EMNLP 2024 Tutorial: https://sites.google.com/view/reasoning-with-explanations
☆11Updated last month
Alternatives and similar repositories for reasoning_with_nle_emnlp_2024
Users that are interested in reasoning_with_nle_emnlp_2024 are comparing it to the libraries listed below
Sorting:
- ☆86Updated last year
- ☆82Updated 2 years ago
- ☆58Updated 5 months ago
- ☆26Updated 2 years ago
- Companion code for FanOutQA: Multi-Hop, Multi-Document Question Answering for Large Language Models (ACL 2024)☆53Updated last week
- Code for the paper "Open Domain Question Answering with A Unified Knowledge Interface" (ACL 2022)☆57Updated last year
- First explanation metric (diagnostic report) for text generation evaluation☆61Updated 2 months ago
- ☆15Updated 2 years ago
- ☆28Updated last year
- ☆105Updated last year
- ☆36Updated 2 years ago
- The dataset and code for PeerSum at EMNLP'23.☆14Updated last year
- Codebase, data and models for the SummaC paper in TACL☆93Updated 3 months ago
- This code accompanies the paper DisentQA: Disentangling Parametric and Contextual Knowledge with Counterfactual Question Answering.☆17Updated 2 years ago
- ☆75Updated last year
- ☆17Updated 4 years ago
- HANNA, a large annotated dataset of Human-ANnotated NArratives for ASG evaluation.☆33Updated 7 months ago
- ☆50Updated last year
- NoMIRACL: A multilingual hallucination evaluation dataset to evaluate LLM robustness in RAG against first-stage retrieval errors on 18 la…☆24Updated 5 months ago
- FRANK: Factuality Evaluation Benchmark☆55Updated 2 years ago
- ☆43Updated 2 years ago
- The Unreliability of Explanations in Few-shot Prompting for Textual Reasoning (NeurIPS 2022)☆15Updated 2 years ago
- Can Large Language Models Be an Alternative to Human Evaluations?☆9Updated last year
- Code and Data for NeurIPS2021 Paper "A Dataset for Answering Time-Sensitive Questions"☆69Updated 3 years ago
- ☆27Updated last year
- The implementation of "RQUGE: Reference-Free Metric for Evaluating Question Generation by Answering the Question" [ACL 2023]☆16Updated last year
- ACL2023 - AlignScore, a metric for factual consistency evaluation.☆127Updated last year
- Easy-to-use framework for evaluating cross-lingual consistency of factual knowledge (Supported LLaMA, BLOOM, mT5, RoBERTa, etc.) Paper he…☆23Updated 2 months ago
- ☆44Updated last year
- Code and models for the paper "Questions Are All You Need to Train a Dense Passage Retriever (TACL 2023)"☆62Updated 2 years ago