hzy312 / Awesome-LLM-Watermark
UP-TO-DATE LLM Watermark paper. 🔥🔥🔥
☆337Updated 4 months ago
Alternatives and similar repositories for Awesome-LLM-Watermark:
Users that are interested in Awesome-LLM-Watermark are comparing it to the libraries listed below
- The lastest paper about detection of LLM-generated text and code☆258Updated 3 months ago
- A survey on harmful fine-tuning attack for large language model☆161Updated last week
- Repository for Towards Codable Watermarking for Large Language Models☆36Updated last year
- Code for watermarking language models☆78Updated 7 months ago
- ☆36Updated 8 months ago
- Repo for SemStamp (NAACL2024) and k-SemStamp (ACL2024)☆20Updated 4 months ago
- ☆571Updated last year
- multi-bit language model watermarking (NAACL 24)☆13Updated 7 months ago
- LLM Unlearning☆153Updated last year
- Accepted by IJCAI-24 Survey Track☆200Updated 8 months ago
- This is the code repository for "Uncovering Safety Risks of Large Language Models through Concept Activation Vector"☆36Updated 5 months ago
- A resource repository for machine unlearning in large language models☆377Updated 3 weeks ago
- Accepted by ECCV 2024☆123Updated 6 months ago
- MarkLLM: An Open-Source Toolkit for LLM Watermarking.(EMNLP 2024 Demo)☆383Updated last month
- A one-stop repository for large language model (LLM) unlearning. Supports TOFU, MUSE and is an easily extensible framework for new datase…☆218Updated last week
- ☆46Updated 10 months ago
- Python package for measuring memorization in LLMs.☆152Updated 5 months ago
- 【ACL 2024】 SALAD benchmark & MD-Judge☆142Updated last month
- ☆14Updated last month
- 😎 up-to-date & curated list of awesome Attacks on Large-Vision-Language-Models papers, methods & resources.☆271Updated 2 weeks ago
- A curated list of LLM Interpretability related material - Tutorial, Library, Survey, Paper, Blog, etc..☆227Updated last month
- [ACL2024-Main] Data and Code for WaterBench: Towards Holistic Evaluation of LLM Watermarks☆24Updated last year
- ☆128Updated 7 months ago
- Official Repository for The Paper: Safety Alignment Should Be Made More Than Just a Few Tokens Deep☆98Updated this week
- Official Repository for ACL 2024 Paper SafeDecoding: Defending against Jailbreak Attacks via Safety-Aware Decoding☆129Updated 9 months ago
- [ICML 2024] Assessing the Brittleness of Safety Alignment via Pruning and Low-Rank Modifications☆76Updated 3 weeks ago
- We jailbreak GPT-3.5 Turbo’s safety guardrails by fine-tuning it on only 10 adversarially designed examples, at a cost of less than $0.20…☆291Updated last year
- ☆18Updated last year
- [ICLR'24] RAIN: Your Language Models Can Align Themselves without Finetuning☆91Updated 11 months ago
- [ICLR 2024] Provable Robust Watermarking for AI-Generated Text☆33Updated last year