Jielin-Qiu / MMWatermark-Robustness
Evaluating Durability: Benchmark Insights into Multimodal Watermarking
☆10Updated 8 months ago
Alternatives and similar repositories for MMWatermark-Robustness:
Users that are interested in MMWatermark-Robustness are comparing it to the libraries listed below
- Preprint: Asymmetry in Low-Rank Adapters of Foundation Models☆34Updated 11 months ago
- ☆13Updated 8 months ago
- Code and data to go with the Zhu et al. paper "An Objective for Nuanced LLM Jailbreaks"☆23Updated 2 months ago
- Certified Patch Robustness via Smoothed Vision Transformers☆42Updated 3 years ago
- ☆17Updated 2 months ago
- Code for Neurips 2024 paper "Shadowcast: Stealthy Data Poisoning Attacks Against Vision-Language Models"☆41Updated last month
- Code for paper "Out-of-Domain Robustness via Targeted Augmentations"☆13Updated last year
- [ECCV 2024] Official PyTorch Implementation of "How Many Unicorns Are in This Image? A Safety Evaluation Benchmark for Vision LLMs"☆75Updated last year
- Respect to the input tensor instead of paramters of NN☆18Updated 2 years ago
- ☆22Updated last week
- Code for the paper "SmoothMix: Training Confidence-calibrated Smoothed Classifiers for Certified Robustness" (NeurIPS 2021)☆21Updated 2 years ago
- Official PyTorch Implementation for Continual Learning and Private Unlearning☆13Updated 2 years ago
- AnyDoor: Test-Time Backdoor Attacks on Multimodal Large Language Models☆49Updated 10 months ago
- Official repository for the paper "Gradient-based Jailbreak Images for Multimodal Fusion Models" (https//arxiv.org/abs/2410.03489)☆13Updated 4 months ago
- [ICML 2023] "Robust Weight Signatures: Gaining Robustness as Easy as Patching Weights?" by Ruisi Cai, Zhenyu Zhang, Zhangyang Wang☆15Updated last year
- [ECCV'24 Oral] The official GitHub page for ''Images are Achilles' Heel of Alignment: Exploiting Visual Vulnerabilities for Jailbreaking …☆22Updated 4 months ago
- Official Implementation of Avoiding spurious correlations via logit correction☆17Updated last year
- Code relative to "Adversarial robustness against multiple and single $l_p$-threat models via quick fine-tuning of robust classifiers"☆18Updated 2 years ago
- Official code for the ICCV2023 paper ``One-bit Flip is All You Need: When Bit-flip Attack Meets Model Training''☆15Updated last year
- Official PyTorch implementation of "CleanCLIP: Mitigating Data Poisoning Attacks in Multimodal Contrastive Learning" @ ICCV 2023☆34Updated last year
- SEAT☆20Updated last year
- ☆27Updated last year
- [ICLR 2023, Spotlight] Indiscriminate Poisoning Attacks on Unsupervised Contrastive Learning☆30Updated last year
- ICLR 2023 paper "Exploring and Exploiting Decision Boundary Dynamics for Adversarial Robustness" by Yuancheng Xu, Yanchao Sun, Micah Gold…☆24Updated last year
- ☆16Updated 9 months ago
- ☆53Updated last year
- Code for Voice Jailbreak Attacks Against GPT-4o.☆28Updated 8 months ago
- Intriguing Properties of Data Attribution on Diffusion Models (ICLR 2024)☆28Updated last year
- The official repository for paper "MLLM-Protector: Ensuring MLLM’s Safety without Hurting Performance"☆33Updated 10 months ago
- Attack AlphaZero Go agents (NeurIPS 2022)☆20Updated 2 years ago