TrustAIResearch / MLHospital
☆43Updated last year
Related projects ⓘ
Alternatives and complementary repositories for MLHospital
- Code Repository for the Paper ---Revisiting the Assumption of Latent Separability for Backdoor Defenses (ICLR 2023)☆34Updated last year
- ☆86Updated 9 months ago
- Official implementation of the CVPR 2022 paper "Backdoor Attacks on Self-Supervised Learning".☆68Updated last year
- Code for ML Doctor☆86Updated 3 months ago
- ☆13Updated 2 years ago
- ☆23Updated 2 years ago
- ☆9Updated 3 years ago
- A curated list of trustworthy Generative AI papers. Daily updating...☆67Updated 2 months ago
- Code for the paper "BadPrompt: Backdoor Attacks on Continuous Prompts"☆36Updated 4 months ago
- Backdoor Safety Tuning (NeurIPS 2023 & 2024 Spotlight)☆24Updated last week
- [NeurIPS 2022] "Randomized Channel Shuffling: Minimal-Overhead Backdoor Attack Detection without Clean Datasets" by Ruisi Cai*, Zhenyu Zh…☆19Updated 2 years ago
- ☆76Updated 3 years ago
- Code&Data for the paper "Watch Out for Your Agents! Investigating Backdoor Threats to LLM-Based Agents" [NeurIPS 2024]☆45Updated last month
- BackdoorLLM: A Comprehensive Benchmark for Backdoor Attacks on Large Language Models☆74Updated 2 months ago
- The official implementation of our pre-print paper "Automatic and Universal Prompt Injection Attacks against Large Language Models".☆34Updated last month
- [ICLR'21] Dataset Inference for Ownership Resolution in Machine Learning☆31Updated 2 years ago
- ☆11Updated 2 years ago
- Code repository for the paper --- [USENIX Security 2023] Towards A Proactive ML Approach for Detecting Backdoor Poison Samples☆22Updated last year
- ☆23Updated last year
- A list of recent papers about adversarial learning☆74Updated this week
- [USENIX'24] Prompt Stealing Attacks Against Text-to-Image Generation Models☆25Updated last month
- Official repository for CVPR'23 paper: Detecting Backdoors in Pre-trained Encoders☆30Updated last year
- [USENIX Security 2025] PoisonedRAG: Knowledge Corruption Attacks to Retrieval-Augmented Generation of Large Language Models☆94Updated last month
- code release for "Unrolling SGD: Understanding Factors Influencing Machine Unlearning" published at EuroS&P'22☆22Updated 2 years ago
- The official implementation of USENIX Security'23 paper "Meta-Sift" -- Ten minutes or less to find a 1000-size or larger clean subset on …☆18Updated last year
- [ICML 2024] COLD-Attack: Jailbreaking LLMs with Stealthiness and Controllability☆110Updated 2 months ago
- Anti-Backdoor learning (NeurIPS 2021)☆78Updated last year
- [ICML 2023] Are Diffusion Models Vulnerable to Membership Inference Attacks?☆31Updated 2 months ago
- Official implementation of "RelaxLoss: Defending Membership Inference Attacks without Losing Utility" (ICLR 2022)☆46Updated 2 years ago
- Codes for NeurIPS 2021 paper "Adversarial Neuron Pruning Purifies Backdoored Deep Models"☆55Updated last year