BrianPulfer / LMWatermarkLinks
Implementation of 'A Watermark for Large Language Models' paper by Kirchenbauer & Geiping et. al.
☆24Updated 2 years ago
Alternatives and similar repositories for LMWatermark
Users that are interested in LMWatermark are comparing it to the libraries listed below
Sorting:
- Code for watermarking language models☆82Updated last year
- Official Implementation of the paper "Three Bricks to Consolidate Watermarks for LLMs"☆48Updated last year
- Source code of paper "An Unforgeable Publicly Verifiable Watermark for Large Language Models" accepted by ICLR 2024☆35Updated last year
- Code and data for paper "A Semantic Invariant Robust Watermark for Large Language Models" accepted by ICLR 2024.☆34Updated 9 months ago
- ☆43Updated 2 years ago
- [EMNLP 2022] Distillation-Resistant Watermarking (DRW) for Model Protection in NLP☆13Updated 2 years ago
- Watermarking Text Generated by Black-Box Language Models☆39Updated last year
- ☆45Updated 6 months ago
- Code of the paper: A Recipe for Watermarking Diffusion Models☆150Updated 9 months ago
- UP-TO-DATE LLM Watermark paper. 🔥🔥🔥☆354Updated 8 months ago
- ☆22Updated 2 years ago
- ☆24Updated 6 months ago
- The official repository of the paper "On the Exploitability of Instruction Tuning".☆64Updated last year
- ☆38Updated last year
- [ICLR 2024] Provable Robust Watermarking for AI-Generated Text☆34Updated last year
- [ICLR 2025] Permute-and-Flip: An optimally robust and watermarkable decoder for LLMs☆19Updated 5 months ago
- This code is the official implementation of WEvade.☆40Updated last year
- ☆27Updated last year
- [ICML 2024] Watermarks in the Sand: Impossibility of Strong Watermarking for Generative Models☆23Updated 11 months ago
- Official repository for "PostMark: A Robust Blackbox Watermark for Large Language Models"☆27Updated last year
- ☆39Updated 2 years ago
- Safe Unlearning: A Surprisingly Effective and Generalizable Solution to Defend Against Jailbreak Attacks☆30Updated last year
- ☆147Updated last year
- Code for the paper "BadPrompt: Backdoor Attacks on Continuous Prompts"☆39Updated last year
- [ICCV 2023] Source code for our paper "Rickrolling the Artist: Injecting Invisible Backdoors into Text-Guided Image Generation Models".☆63Updated last year
- ☆315Updated last year
- Code for our paper "Defending ChatGPT against Jailbreak Attack via Self-Reminder" in NMI.☆53Updated last year
- NeurIPS'24 - LLM Safety Landscape☆28Updated 6 months ago
- ☆162Updated last month
- Code for our S&P'21 paper: Adversarial Watermarking Transformer: Towards Tracing Text Provenance with Data Hiding☆53Updated 2 years ago