☆14Mar 9, 2025Updated 11 months ago
Alternatives and similar repositories for Secure_LLM_System
Users that are interested in Secure_LLM_System are comparing it to the libraries listed below
Sorting:
- A research workbench for developing and testing attacks against large language models, with a focus on prompt injection vulnerabilities a…☆38Updated this week
- ☆29Oct 23, 2024Updated last year
- ☆27Sep 11, 2025Updated 5 months ago
- PFI: Prompt Flow Integrity to Prevent Privilege Escalation in LLM Agents☆26Mar 26, 2025Updated 11 months ago
- Distribution Preserving Backdoor Attack in Self-supervised Learning☆20Jan 27, 2024Updated 2 years ago
- Code repo for the paper: Attacking Vision-Language Computer Agents via Pop-ups☆50Dec 23, 2024Updated last year
- ☆23Sep 15, 2024Updated last year
- A curated list of awesome resources about LLM supply chain security (including papers, security reports and CVEs)☆96Jan 20, 2025Updated last year
- ☆52Feb 8, 2025Updated last year
- [NeurIPS 2022] "Randomized Channel Shuffling: Minimal-Overhead Backdoor Attack Detection without Clean Datasets" by Ruisi Cai*, Zhenyu Zh…☆21Oct 1, 2022Updated 3 years ago
- This is the repository that introduces research topics related to protecting intellectual property (IP) of AI from a data-centric perspec…☆23Oct 30, 2023Updated 2 years ago
- Codebase for Obfuscated Activations Bypass LLM Latent-Space Defenses☆29Feb 11, 2025Updated last year
- Code to generate NeuralExecs (prompt injection for LLMs)☆27Oct 5, 2025Updated 4 months ago
- An implementation for the paper "A Little Is Enough: Circumventing Defenses For Distributed Learning" (NeurIPS 2019)☆28Jun 29, 2023Updated 2 years ago
- ☆29Aug 31, 2025Updated 6 months ago
- This is the source code for MEA-Defender. Our paper is accepted by the IEEE Symposium on Security and Privacy (S&P) 2024.☆29Nov 19, 2023Updated 2 years ago
- Code for Voice Jailbreak Attacks Against GPT-4o.☆36May 31, 2024Updated last year
- ☆27Nov 9, 2022Updated 3 years ago
- Repo for the research paper "SecAlign: Defending Against Prompt Injection with Preference Optimization"☆86Jul 24, 2025Updated 7 months ago
- 北京邮电大学信通院C++上机题☆14Feb 20, 2021Updated 5 years ago
- ☆11Dec 23, 2024Updated last year
- Agent Security Bench (ASB)☆186Oct 27, 2025Updated 4 months ago
- Bad Characters: Imperceptible NLP Attacks☆35Apr 9, 2024Updated last year
- ☆12May 6, 2022Updated 3 years ago
- This repository contains the source code for "Membership Inference Attacks as Privacy Tools: Reliability, Disparity and Ensemble", In Pro…☆10Jan 2, 2026Updated last month
- ☆37Sep 30, 2024Updated last year
- ☆33Aug 24, 2023Updated 2 years ago
- Flow Integrity Deterministic Enforcement System. Mechanisms for securing AI agents with information-flow control.☆75May 30, 2025Updated 8 months ago
- Code for the paper "Defeating Prompt Injections by Design"☆262Jun 20, 2025Updated 8 months ago
- Text-CRS: A Generalized Certified Robustness Framework against Textual Adversarial Attacks (IEEE S&P 2024)☆34Jun 29, 2025Updated 7 months ago
- [EMNLP 2025 Oral] IPIGuard: A Novel Tool Dependency Graph-Based Defense Against Indirect Prompt Injection in LLM Agents☆16Sep 16, 2025Updated 5 months ago
- On the Robustness of GUI Grounding Models Against Image Attacks☆12Apr 8, 2025Updated 10 months ago
- ☆24Jul 27, 2024Updated last year
- BrainWash: A Poisoning Attack to Forget in Continual Learning☆12Apr 15, 2024Updated last year
- [NeurIPS 2025] The official implementation of the paper "DRIFT: Dynamic Rule-Based Defense with Injection Isolation for Securing LLM Agen…☆39Feb 14, 2026Updated last week
- 中国农业大学计算机图形学完整代码和报告,欢迎参考学习☆10Dec 12, 2021Updated 4 years ago
- ☆117Jul 2, 2024Updated last year
- ☆48Sep 29, 2024Updated last year
- [ICLR 2022] Boosting Randomized Smoothing with Variance Reduced Classifiers☆12Mar 29, 2022Updated 3 years ago