Lingzhi-WANG / KGAUnlearnLinks
☆19Updated 2 years ago
Alternatives and similar repositories for KGAUnlearn
Users that are interested in KGAUnlearn are comparing it to the libraries listed below
Sorting:
- [NeurIPS23 (Spotlight)] "Model Sparsity Can Simplify Machine Unlearning" by Jinghan Jia*, Jiancheng Liu*, Parikshit Ram, Yuguang Yao, Gao…☆81Updated last year
- LLM Unlearning☆178Updated 2 years ago
- ☆77Updated 3 years ago
- "In-Context Unlearning: Language Models as Few Shot Unlearners". Martin Pawelczyk, Seth Neel* and Himabindu Lakkaraju*; ICML 2024.☆28Updated 2 years ago
- ☆58Updated 5 years ago
- LAMP: Extracting Text from Gradients with Language Model Priors (NeurIPS '22)☆27Updated 6 months ago
- A survey of privacy problems in Large Language Models (LLMs). Contains summary of the corresponding paper along with relevant code☆68Updated last year
- A toolkit to assess data privacy in LLMs (under development)☆65Updated 11 months ago
- Code for paper "Universal Jailbreak Backdoors from Poisoned Human Feedback"☆66Updated last year
- ☆48Updated 10 months ago
- [ICLR 2025] A Closer Look at Machine Unlearning for Large Language Models☆41Updated last year
- A survey on harmful fine-tuning attack for large language model☆225Updated last month
- Backdoor Safety Tuning (NeurIPS 2023 & 2024 Spotlight)☆27Updated last year
- This is the official code for the paper "Vaccine: Perturbation-aware Alignment for Large Language Models" (NeurIPS2024)☆48Updated last year
- [AAAI, ICLR TP] Fast Machine Unlearning Without Retraining Through Selective Synaptic Dampening☆55Updated last year
- ☆31Updated 9 months ago
- Python package for measuring memorization in LLMs.☆175Updated 5 months ago
- Code and data for paper "A Semantic Invariant Robust Watermark for Large Language Models" accepted by ICLR 2024.☆37Updated last year
- ☆39Updated last year
- ICLR2024 Paper. Showing properties of safety tuning and exaggerated safety.☆89Updated last year
- ☆28Updated this week
- This is the official code for the paper "Safety Tax: Safety Alignment Makes Your Large Reasoning Models Less Reasonable".☆26Updated 9 months ago
- [CCS-LAMPS'24] LLM IP Protection Against Model Merging☆16Updated last year
- ☆38Updated 2 years ago
- [EMNLP 2023] Poisoning Retrieval Corpora by Injecting Adversarial Passages https://arxiv.org/abs/2310.19156☆43Updated 2 years ago
- ☆114Updated 2 years ago
- Code for watermarking language models☆84Updated last year
- ☆47Updated last year
- Reconstructive Neuron Pruning for Backdoor Defense (ICML 2023)☆39Updated last year
- ☆70Updated 10 months ago