YichenZW / Robust-Det
The code implementation of the paper Stumbling Blocks: Stress Testing the Robustness of Machine-Generated Text Detectors Under Attacks (ACL 2024 main) by Yichen Wang, Shangbin Feng, Abe Bohan Hou, Xiao Pu, Chao Shen, Xiaoming Liu, and Yulia Tsvetkov, and Tianxing He, mainly at Paul G. Allen School of CSE, University of Washington.
☆11Updated 6 months ago
Alternatives and similar repositories for Robust-Det:
Users that are interested in Robust-Det are comparing it to the libraries listed below
- AbstainQA, ACL 2024☆25Updated 3 months ago
- Github repository for "FELM: Benchmarking Factuality Evaluation of Large Language Models" (NeurIPS 2023)☆57Updated last year
- ☆57Updated 3 weeks ago
- ☆19Updated last year
- ☆29Updated 8 months ago
- Recent papers on (1) Psychology of LLMs; (2) Biases in LLMs.☆44Updated last year
- ☆24Updated 3 months ago
- ☆38Updated last year
- ✨ Resolving Knowledge Conflicts in Large Language Models, COLM 2024☆15Updated 3 months ago
- Semi-Parametric Editing with a Retrieval-Augmented Counterfactual Model☆66Updated 2 years ago
- Safety-J: Evaluating Safety with Critique☆16Updated 5 months ago
- Code associated with Tuning Language Models by Proxy (Liu et al., 2024)☆102Updated 9 months ago
- Steering Knowledge Selection Behaviours in LLMs via SAE-Based Representation Engineering☆39Updated last month
- Evaluating the Ripple Effects of Knowledge Editing in Language Models☆53Updated 9 months ago
- Evaluate the Quality of Critique☆35Updated 7 months ago
- ☆28Updated last year
- Official code implementation of SKU, Accepted by ACL 2024 Findings☆13Updated last month
- Official Code Repository for LM-Steer Paper: "Word Embeddings Are Steers for Language Models" (ACL 2024 Outstanding Paper Award)☆73Updated 3 months ago
- [EMNLP 2024 Findings] To Forget or Not? Towards Practical Knowledge Unlearning for Large Language Models☆25Updated 2 months ago
- Code for ACL 2023 paper "BOLT: Fast Energy-based Controlled Text Generation with Tunable Biases".☆21Updated last year
- [ICML 2023] Code for our paper “Compositional Exemplars for In-context Learning”.☆95Updated last year
- Code & Data for our Paper "Alleviating Hallucinations of Large Language Models through Induced Hallucinations"☆62Updated 10 months ago
- Code for the ACL 2023 Paper "Fact-Checking Complex Claims with Program-Guided Reasoning"☆50Updated last year
- In-Context Sharpness as Alerts: An Inner Representation Perspective for Hallucination Mitigation (ICML 2024)☆50Updated 9 months ago
- LoFiT: Localized Fine-tuning on LLM Representations☆30Updated this week
- Code for the ACL-2022 paper "Knowledge Neurons in Pretrained Transformers"☆162Updated 8 months ago
- ☆44Updated 4 months ago
- ☆34Updated 2 months ago
- Repo for paper: Examining LLMs' Uncertainty Expression Towards Questions Outside Parametric Knowledge☆12Updated 10 months ago
- [EMNLP 2023] MQuAKE: Assessing Knowledge Editing in Language Models via Multi-Hop Questions☆105Updated 4 months ago