chrisliu298 / awesome-llm-unlearningLinks
A resource repository for machine unlearning in large language models
☆516Updated 2 weeks ago
Alternatives and similar repositories for awesome-llm-unlearning
Users that are interested in awesome-llm-unlearning are comparing it to the libraries listed below
Sorting:
- [NeurIPS D&B '25] The one-stop repository for LLM unlearning☆453Updated last week
- A survey on harmful fine-tuning attack for large language model☆229Updated last month
- UP-TO-DATE LLM Watermark paper. 🔥🔥🔥☆370Updated last year
- ☆178Updated last month
- LLM Unlearning☆178Updated 2 years ago
- A resource repository for representation engineering in large language models☆145Updated last year
- A curated list of LLM Interpretability related material - Tutorial, Library, Survey, Paper, Blog, etc..☆290Updated last week
- ☆28Updated 2 weeks ago
- Official Repository for The Paper: Safety Alignment Should Be Made More Than Just a Few Tokens Deep☆166Updated 8 months ago
- Toolkit for evaluating the trustworthiness of generative foundation models.☆124Updated 4 months ago
- A toolkit to assess data privacy in LLMs (under development)☆66Updated last year
- Python package for measuring memorization in LLMs.☆177Updated 5 months ago
- Accepted by IJCAI-24 Survey Track☆225Updated last year
- [ICML 2024] TrustLLM: Trustworthiness in Large Language Models☆619Updated 6 months ago
- Awesome-LLM-Robustness: a curated list of Uncertainty, Reliability and Robustness in Large Language Models☆802Updated 7 months ago
- The lastest paper about detection of LLM-generated text and code☆281Updated 6 months ago
- [USENIX Security 2025] PoisonedRAG: Knowledge Corruption Attacks to Retrieval-Augmented Generation of Large Language Models☆222Updated last month
- We jailbreak GPT-3.5 Turbo’s safety guardrails by fine-tuning it on only 10 adversarially designed examples, at a cost of less than $0.20…☆335Updated last year
- awesome SAE papers☆69Updated 7 months ago
- A curated list of resources for activation engineering☆120Updated 2 months ago
- A survey of privacy problems in Large Language Models (LLMs). Contains summary of the corresponding paper along with relevant code☆68Updated last year
- Accepted by ECCV 2024☆179Updated last year
- Papers and resources related to the security and privacy of LLMs 🤖☆554Updated 6 months ago
- ☆156Updated 2 years ago
- ☆70Updated 10 months ago
- ☆61Updated 7 months ago
- ☆55Updated last year
- Model Merging in LLMs, MLLMs, and Beyond: Methods, Theories, Applications and Opportunities. ACM Computing Surveys, 2025.☆630Updated this week
- [ICML 2024] Assessing the Brittleness of Safety Alignment via Pruning and Low-Rank Modifications☆88Updated 9 months ago
- Must-read Papers on Knowledge Editing for Large Language Models.☆1,210Updated 5 months ago