liujch1998 / veraLinks
☆16Updated 2 years ago
Alternatives and similar repositories for vera
Users that are interested in vera are comparing it to the libraries listed below
Sorting:
- Rationale-enhanced language models are better continual relation learners (EMNLP 2023 Main Conference)☆12Updated 2 years ago
- Visual and Embodied Concepts evaluation benchmark☆21Updated 2 years ago
- Source code and data used in the papers ViQuAE (Lerner et al., SIGIR'22), Multimodal ICT (Lerner et al., ECIR'23) and Cross-modal Retriev…☆38Updated last year
- ☆19Updated 2 years ago
- ☆25Updated 2 years ago
- [EMNLP'2023 Findings] MoqaGPT, for zero-shot multimodal question answering with LLMs☆13Updated last year
- Code & Data for our Paper "Alleviating Hallucinations of Large Language Models through Induced Hallucinations"☆69Updated last year
- ☆43Updated 2 years ago
- ☆149Updated 3 years ago
- Findings of EMNLP 2023: InfoCL: Alleviating Catastrophic Forgetting in Continual Text Classification from An Information Theoretic Perspe…☆14Updated last year
- EMNLP2023 - InfoSeek: A New VQA Benchmark focus on Visual Info-Seeking Questions☆25Updated last year
- Dialogue Planning via Brownian Bridge Stochastic Process for Goal-directed Proactive Dialogue (ACL Findings 2023)☆21Updated 2 months ago
- Active Example Selection for In-Context Learning (EMNLP'22)☆49Updated last year
- ☆62Updated last year
- ☆16Updated 3 years ago
- ☆68Updated 2 years ago
- ☆57Updated last year
- Source code for Truth-Aware Context Selection: Mitigating the Hallucinations of Large Language Models Being Misled by Untruthful Contexts☆17Updated last year
- [EMNLP 2022] TaCube: Pre-computing Data Cubes for Answering Numerical-Reasoning Questions over Tabular Data☆17Updated 2 years ago
- Code and data for the paper: On the Reliability of Psychological Scales on Large Language Models☆30Updated last month
- ☆32Updated 2 years ago
- ☆64Updated 3 years ago
- ACL'2023: Multi-Task Pre-Training of Modular Prompt for Few-Shot Learning☆40Updated 3 years ago
- BeHonest: Benchmarking Honesty in Large Language Models☆34Updated last year
- ☆27Updated 2 years ago
- [EMNLP 2023] MQuAKE: Assessing Knowledge Editing in Language Models via Multi-Hop Questions☆118Updated last year
- ☆12Updated last year
- Github repository for "FELM: Benchmarking Factuality Evaluation of Large Language Models" (NeurIPS 2023)☆63Updated 2 years ago
- [ICLR'24 Spotlight] "Adaptive Chameleon or Stubborn Sloth: Revealing the Behavior of Large Language Models in Knowledge Conflicts"☆81Updated last year
- ☆15Updated 3 years ago