Alrope123 / prompt-waywardness
☆13Updated 2 years ago
Related projects ⓘ
Alternatives and complementary repositories for prompt-waywardness
- ☆78Updated 2 years ago
- TBC☆26Updated 2 years ago
- ☆16Updated last year
- Repository for ACL 2022 paper Mix and Match: Learning-free Controllable Text Generation using Energy Language Models☆41Updated 2 years ago
- ☆48Updated last year
- Findings of ACL'2023: Optimizing Test-Time Query Representations for Dense Retrieval☆29Updated last year
- ☆42Updated 9 months ago
- ☆57Updated 2 years ago
- ☆74Updated last year
- Influence Experiments☆35Updated last year
- EMNLP 2022: Generating Natural Language Proofs with Verifier-Guided Search https://arxiv.org/abs/2205.12443☆81Updated last month
- Github repository for "FELM: Benchmarking Factuality Evaluation of Large Language Models" (NeurIPS 2023)☆54Updated 10 months ago
- EMNLP 2022: "MABEL: Attenuating Gender Bias using Textual Entailment Data" https://arxiv.org/abs/2210.14975☆37Updated 11 months ago
- ☆48Updated last year
- This repository contains the dataset and code for "WiCE: Real-World Entailment for Claims in Wikipedia" in EMNLP 2023.☆39Updated 10 months ago
- [EMNLP 2022] TemporalWiki: A Lifelong Benchmark for Training and Evaluating Ever-Evolving Language Models☆66Updated 5 months ago
- ☆37Updated 7 months ago
- NAACL 2022: Can Rationalization Improve Robustness? https://arxiv.org/abs/2204.11790☆27Updated last year
- Restore safety in fine-tuned language models through task arithmetic☆26Updated 7 months ago
- NumGLUE: A Suite of Fundamental yet Challenging Mathematical Reasoning Tasks☆20Updated 2 years ago
- This repository accompanies our paper “Do Prompt-Based Models Really Understand the Meaning of Their Prompts?”☆84Updated 2 years ago
- ☆80Updated last year
- Code for preprint: Summarizing Differences between Text Distributions with Natural Language☆42Updated last year
- [ICML 2023] Exploring the Benefits of Training Expert Language Models over Instruction Tuning☆97Updated last year
- [NeurIPS 2022] Generating Training Data with Language Models: Towards Zero-Shot Language Understanding☆62Updated 2 years ago
- Code and Data for NeurIPS2021 Paper "A Dataset for Answering Time-Sensitive Questions"☆62Updated 2 years ago
- ☆33Updated 2 years ago
- WikiWhy is a new benchmark for evaluating LLMs' ability to explain between cause-effect relationships. It is a QA dataset containing 9000…☆46Updated 11 months ago
- [EMNLP-2022 Findings] Code for paper “ProGen: Progressive Zero-shot Dataset Generation via In-context Feedback”.☆24Updated last year
- The official code of TACL 2021, "Did Aristotle Use a Laptop? A Question Answering Benchmark with Implicit Reasoning Strategies".☆64Updated 2 years ago