voidism / EAR
Code for the ACL 2023 long paper - Expand, Rerank, and Retrieve: Query Reranking for Open-Domain Question Answering
☆35Updated last year
Alternatives and similar repositories for EAR:
Users that are interested in EAR are comparing it to the libraries listed below
- Code and data for paper "Context-faithful Prompting for Large Language Models".☆39Updated 2 years ago
- [ACL'24 Oral] Analysing The Impact of Sequence Composition on Language Model Pre-Training☆20Updated 7 months ago
- Token-level Reference-free Hallucination Detection☆94Updated last year
- "FiD-ICL: A Fusion-in-Decoder Approach for Efficient In-Context Learning" (ACL 2023)☆13Updated last year
- Open-WikiTable :Dataset for Open Domain Question Answering with Complex Reasoning over Table☆23Updated last year
- [ICLR'24 Spotlight] "Adaptive Chameleon or Stubborn Sloth: Revealing the Behavior of Large Language Models in Knowledge Conflicts"☆67Updated 11 months ago
- Repository for "Propagating Knowledge Updates to LMs Through Distillation" (NeurIPS 2023).☆25Updated 7 months ago
- The code implementation of the EMNLP2022 paper: DisCup: Discriminator Cooperative Unlikelihood Prompt-tuning for Controllable Text Gene…☆26Updated last year
- Easy-to-use MIRAGE code for faithful answer attribution in RAG applications. Paper: https://aclanthology.org/2024.emnlp-main.347/☆21Updated 3 weeks ago
- Github repository for "FELM: Benchmarking Factuality Evaluation of Large Language Models" (NeurIPS 2023)☆58Updated last year
- Merging Generated and Retrieved Knowledge for Open-Domain QA (EMNLP 2023)☆22Updated last year
- Repo for Llatrieval☆29Updated 7 months ago
- First explanation metric (diagnostic report) for text generation evaluation☆62Updated 3 weeks ago
- This is the code for the Submission 3358 at NeurIPS 2022.☆21Updated 2 years ago
- Implementation of the paper: "Making Retrieval-Augmented Language Models Robust to Irrelevant Context"☆65Updated 7 months ago
- ☆86Updated last year
- Contrastive Chain-of-Thought Prompting☆58Updated last year
- ☆52Updated 7 months ago
- Evaluate the Quality of Critique☆34Updated 9 months ago
- Technical Report: Is ChatGPT a Good NLG Evaluator? A Preliminary Study☆43Updated 2 years ago
- ☆44Updated last year
- ☆48Updated 11 months ago
- ☆85Updated 2 years ago
- The project page for "SCITAB: A Challenging Benchmark for Compositional Reasoning and Claim Verification on Scientific Tables"☆20Updated last year
- This repository contains the dataset and code for "WiCE: Real-World Entailment for Claims in Wikipedia" in EMNLP 2023.☆41Updated last year
- BRIGHT: A Realistic and Challenging Benchmark for Reasoning-Intensive Retrieval☆92Updated last month
- WikiWhy is a new benchmark for evaluating LLMs' ability to explain between cause-effect relationships. It is a QA dataset containing 9000…☆47Updated last year
- Code and resources for papers "Generation-Augmented Retrieval for Open-Domain Question Answering" and "Reader-Guided Passage Reranking fo…☆74Updated 3 years ago
- IntructIR, a novel benchmark specifically designed to evaluate the instruction following ability in information retrieval models. Our foc…☆31Updated 9 months ago
- Companion code for FanOutQA: Multi-Hop, Multi-Document Question Answering for Large Language Models (ACL 2024)☆51Updated 3 weeks ago