myZeratul / Causal-DebiasLinks
☆10Updated 2 years ago
Alternatives and similar repositories for Causal-Debias
Users that are interested in Causal-Debias are comparing it to the libraries listed below
Sorting:
- ☆27Updated last year
- This is the repo for the survey of Bias and Fairness in IR with LLMs.☆53Updated 2 months ago
- ☆56Updated 3 months ago
- Implementation of the paper "Exploring the Universal Vulnerability of Prompt-based Learning Paradigm" on Findings of NAACL 2022☆29Updated 2 years ago
- ☆20Updated 2 months ago
- ☆26Updated 6 months ago
- [ACL2023] Preserving Commonsense Knowledge from Pre-trained Language Models via Causal Inference☆23Updated last year
- ☆12Updated 9 months ago
- code of paper "Defending Against Alignment-Breaking Attacks via Robustly Aligned LLM"☆12Updated last year
- ☆24Updated last year
- Official Code for ACL 2023 paper: "Ethicist: Targeted Training Data Extraction Through Loss Smoothed Soft Prompting and Calibrated Confid…☆23Updated 2 years ago
- Official implementation of the EMNLP 2021 paper "ONION: A Simple and Effective Defense Against Textual Backdoor Attacks"☆33Updated 3 years ago
- The dataset and code for the ICLR 2024 paper "Can LLM-Generated Misinformation Be Detected?"☆66Updated 6 months ago
- [ACL 2024] Shifting Attention to Relevance: Towards the Predictive Uncertainty Quantification of Free-Form Large Language Models☆49Updated 9 months ago
- "In-Context Unlearning: Language Models as Few Shot Unlearners". Martin Pawelczyk, Seth Neel* and Himabindu Lakkaraju*; ICML 2024.☆26Updated last year
- ☆28Updated 11 months ago
- Data for our paper "Defending ChatGPT against Jailbreak Attack via Self-Reminder"☆18Updated last year
- [ACL2024-Main] Data and Code for WaterBench: Towards Holistic Evaluation of LLM Watermarks☆26Updated last year
- Code for the paper "BadPrompt: Backdoor Attacks on Continuous Prompts"☆36Updated 10 months ago
- Official implementation of our paper "Separate the Wheat from the Chaff: Model Deficiency Unlearning via Parameter-Efficient Module Opera…☆11Updated 8 months ago
- Chain of Attack: a Semantic-Driven Contextual Multi-Turn attacker for LLM☆31Updated 4 months ago
- ☆18Updated 8 months ago
- FedJudge: Federated Legal Large Language Model☆34Updated 8 months ago
- What does the bot say? ACL 2024☆22Updated 9 months ago
- ☆41Updated 8 months ago
- [ACL 2024] Defending Large Language Models Against Jailbreaking Attacks Through Goal Prioritization☆24Updated 10 months ago
- Code for the paper "Be Careful about Poisoned Word Embeddings: Exploring the Vulnerability of the Embedding Layers in NLP Models" (NAACL-…☆41Updated 3 years ago
- WWW 2024: New Frontiers of Knowledge Graph Reasoning: Recent Advances and Future Trends☆17Updated last year
- A Task of Fictitious Unlearning for VLMs☆17Updated 2 months ago
- Code for Findings-EMNLP 2023 paper: Multi-step Jailbreaking Privacy Attacks on ChatGPT☆33Updated last year