jzhang538 / BadMerging
[CCS 2024] "BadMerging: Backdoor Attacks Against Model Merging": official code implementation.
☆27Updated 7 months ago
Alternatives and similar repositories for BadMerging:
Users that are interested in BadMerging are comparing it to the libraries listed below
- (CVPR 2025) Official implementation to DELT: A Simple Diversity-driven EarlyLate Training for Dataset Distillation which outperforms SOTA…☆20Updated last month
- Improving Your Model Ranking on Chatbot Arena by Vote Rigging☆19Updated last month
- ☆27Updated last year
- [TMLR'24] This repository includes the official implementation our paper "FedConv: Enhancing Convolutional Neural Networks for Handling D…☆25Updated 11 months ago
- Edit Away and My Face Will not Stay: Personal Biometric Defense against Malicious Generative Editing☆32Updated 3 months ago
- Code for Neurips 2024 paper "Shadowcast: Stealthy Data Poisoning Attacks Against Vision-Language Models"☆44Updated 2 months ago
- Unsolvable Problem Detection: Evaluating Trustworthiness of Vision Language Models☆75Updated 6 months ago
- The official implementation of ECCV'24 paper "To Generate or Not? Safety-Driven Unlearned Diffusion Models Are Still Easy To Generate Uns…☆72Updated last month
- [ECCV 2024] The official code for "AdaShield: Safeguarding Multimodal Large Language Models from Structure-based Attack via Adaptive Shi…☆55Updated 8 months ago
- A Simple Baseline Achieving Over 90% Success Rate Against the Strong Black-box Models of GPT-4.5/4o/o1. Paper at: https://arxiv.org/abs/2…☆46Updated last week
- This repository contains the source code, datasets, and scripts for the paper "GenderCARE: A Comprehensive Framework for Assessing and Re…☆20Updated 7 months ago
- [CCS 2024] Optimization-based Prompt Injection Attack to LLM-as-a-Judge☆15Updated 4 months ago
- [ICLR 2024] Inducing High Energy-Latency of Large Vision-Language Models with Verbose Images☆32Updated last year
- ☆23Updated 3 weeks ago
- ☆26Updated 3 months ago
- [ICML 2024] Agent Smith: A Single Image Can Jailbreak One Million Multimodal LLM Agents Exponentially Fast☆98Updated last year
- A novel approach to improve the safety of large language models, enabling them to transition effectively from unsafe to safe state.☆58Updated 2 months ago
- Official code for the ICCV2023 paper ``One-bit Flip is All You Need: When Bit-flip Attack Meets Model Training''☆18Updated last year
- [ICLR 2025] Dissecting Adversarial Robustness of Multimodal LM Agents☆77Updated last month
- AnyDoor: Test-Time Backdoor Attacks on Multimodal Large Language Models☆53Updated 11 months ago
- List of T2I safety papers, updated daily, welcome to discuss using Discussions☆59Updated 7 months ago
- [ICML 2024] Unsupervised Adversarial Fine-Tuning of Vision Embeddings for Robust Large Vision-Language Models☆124Updated 4 months ago
- [ICML 2024] Prompting4Debugging: Red-Teaming Text-to-Image Diffusion Models by Finding Problematic Prompts (Official Pytorch Implementati…☆41Updated 4 months ago
- [ECCV'24 Oral] The official GitHub page for ''Images are Achilles' Heel of Alignment: Exploiting Visual Vulnerabilities for Jailbreaking …☆28Updated 5 months ago
- [NeurIPS 2024] Official implementation of the paper “Ferrari: Federated Feature Unlearning via Optimizing Feature Sensitivity"☆11Updated 2 weeks ago
- An official implementation of "Catastrophic Failure of LLM Unlearning via Quantization" (ICLR 2025)☆26Updated last month
- This is a collection of awesome papers I have read (carefully or roughly) in the fields of security in diffusion models. Any suggestions …☆24Updated 4 months ago
- ☆10Updated 3 months ago
- Official Repository of Personalized Visual Instruct Tuning☆28Updated 3 weeks ago
- The official implementation for "Towards Physically Realizable Adversarial Attacks in Embodied Vision Navigation"☆19Updated 3 weeks ago