☆77Dec 19, 2024Updated last year
Alternatives and similar repositories for PLeak
Users that are interested in PLeak are comparing it to the libraries listed below
Sorting:
- ☆118Jul 2, 2024Updated last year
- ☆23Jan 17, 2025Updated last year
- Consuming Resrouce via Auto-generation for LLM-DoS Attack under Black-box Settings☆18Sep 1, 2025Updated 6 months ago
- Code&Data for the paper "Watch Out for Your Agents! Investigating Backdoor Threats to LLM-Based Agents" [NeurIPS 2024]☆109Sep 27, 2024Updated last year
- ☆12May 6, 2022Updated 3 years ago
- AI fun☆27Feb 27, 2025Updated last year
- Official repo for GPTFUZZER : Red Teaming Large Language Models with Auto-Generated Jailbreak Prompts☆568Updated this week
- Code repo for the paper: Attacking Vision-Language Computer Agents via Pop-ups☆51Dec 23, 2024Updated last year
- ☆12Dec 22, 2025Updated 2 months ago
- ☆75Jan 21, 2026Updated last month
- A package that achieves 95%+ transfer attack success rate against GPT-4☆26Oct 24, 2024Updated last year
- A fast + lightweight implementation of the GCG algorithm in PyTorch☆318May 13, 2025Updated 9 months ago
- The repo for paper: Exploiting the Index Gradients for Optimization-Based Jailbreaking on Large Language Models.☆13Dec 16, 2024Updated last year
- Official Code Implementation for the CCS 2022 Paper "On the Privacy Risks of Cell-Based NAS Architectures"☆11Nov 21, 2022Updated 3 years ago
- ☆14Mar 9, 2025Updated 11 months ago
- ☆14Feb 26, 2025Updated last year
- LobotoMl is a set of scripts and tools to assess production deployments of ML services☆10May 16, 2022Updated 3 years ago
- [ICCV-2025] Universal Adversarial Attack, Multimodal Adversarial Attacks, VLP models, Contrastive Learning, Cross-modal Perturbation Gene…☆35Jul 10, 2025Updated 7 months ago
- ☆78May 28, 2022Updated 3 years ago
- Codes for paper "SafeAgentBench: A Benchmark for Safe Task Planning of \\ Embodied LLM Agents"☆65Feb 25, 2025Updated last year
- Unofficial Iranian hackers group disk wiper malware aka "Shamoon" in .NET 2.0☆13Dec 23, 2018Updated 7 years ago
- ☆13Jun 15, 2024Updated last year
- [USENIX Security 2025] SOFT: Selective Data Obfuscation for Protecting LLM Fine-tuning against Membership Inference Attacks☆20Sep 18, 2025Updated 5 months ago
- Code for Voice Jailbreak Attacks Against GPT-4o.☆36May 31, 2024Updated last year
- ☆34Aug 28, 2024Updated last year
- Fine-tuning base models to build robust task-specific models☆34Apr 11, 2024Updated last year
- Generating API Parameter Security Rules with LLM for API Misuse Detection (Accepted by NDSS 2025)☆18Mar 5, 2025Updated 11 months ago
- AI Security Research☆15Jun 21, 2023Updated 2 years ago
- Official implementation of "Data Mixture Inference: What do BPE tokenizers reveal about their training data?"☆18May 15, 2025Updated 9 months ago
- ☆39May 17, 2025Updated 9 months ago
- ☆86Mar 20, 2025Updated 11 months ago
- [EMNLP 2025] Reasoning-to-Defend: Safety-Aware Reasoning Can Defend Large Language Models from Jailbreaking☆12Aug 22, 2025Updated 6 months ago
- ☆18Oct 12, 2022Updated 3 years ago
- ☆37Oct 17, 2024Updated last year
- Papers and resources related to the security and privacy of LLMs 🤖☆568Jun 8, 2025Updated 8 months ago
- [ICLR 2025] Dissecting adversarial robustness of multimodal language model agents☆124Feb 19, 2025Updated last year
- ☆13Dec 8, 2022Updated 3 years ago
- [NeurIPS 2025] BackdoorLLM: A Comprehensive Benchmark for Backdoor Attacks and Defenses on Large Language Models☆276Feb 2, 2026Updated last month
- Distribution Preserving Backdoor Attack in Self-supervised Learning☆20Jan 27, 2024Updated 2 years ago