Reapor-Yurnero / imprompter
Codebase of https://arxiv.org/abs/2410.14923
☆35Updated 2 months ago
Alternatives and similar repositories for imprompter:
Users that are interested in imprompter are comparing it to the libraries listed below
- ☆26Updated 2 months ago
- Project Mantis: Hacking Back the AI-Hacker; Prompt Injection as a Defense Against LLM-driven Cyberattacks☆59Updated last month
- TaskTracker is an approach to detecting task drift in Large Language Models (LLMs) by analysing their internal activations. It provides a…☆42Updated last month
- Red-Teaming Language Models with DSPy☆153Updated 9 months ago
- A prompt injection game to collect data for robust ML research☆49Updated 3 weeks ago
- A subset of jailbreaks automatically discovered by the Haize Labs haizing suite.☆88Updated 7 months ago
- A benchmark for evaluating the robustness of LLMs and defenses to indirect prompt injection attacks.☆53Updated 9 months ago
- ☆67Updated last month
- Risks and targets for assessing LLMs & LLM vulnerabilities☆30Updated 7 months ago
- ☆45Updated last month
- ☆48Updated 3 months ago
- ☆63Updated 3 months ago
- ComPromptMized: Unleashing Zero-click Worms that Target GenAI-Powered Applications☆197Updated 10 months ago
- ☆27Updated last month
- Jailbreaking Leading Safety-Aligned LLMs with Simple Adaptive Attacks [arXiv, Apr 2024]☆247Updated 3 months ago
- SecGPT: An execution isolation architecture for LLM-based systems☆57Updated last month
- Contains random samples referenced in the paper "Sleeper Agents: Training Robustly Deceptive LLMs that Persist Through Safety Training".☆92Updated 10 months ago
- A benchmark for prompt injection detection systems.☆94Updated 4 months ago
- Curation of prompts that are known to be adversarial to large language models☆177Updated last year
- The official implementation of our pre-print paper "Automatic and Universal Prompt Injection Attacks against Large Language Models".☆39Updated 2 months ago
- Package to optimize Adversarial Attacks against (Large) Language Models with Varied Objectives☆66Updated 10 months ago
- Machine Learning Attack Series☆57Updated 8 months ago
- A productionized greedy coordinate gradient (GCG) attack tool for large language models (LLMs)☆75Updated last month
- Repo for the research paper "Aligning LLMs to Be Robust Against Prompt Injection"☆32Updated last month
- ☆114Updated last month
- Whispers in the Machine: Confidentiality in LLM-integrated Systems☆31Updated last month
- Do you want to learn AI Security but don't know where to start ? Take a look at this map.☆20Updated 8 months ago
- Supply chain security for ML☆122Updated this week
- A Dynamic Environment to Evaluate Attacks and Defenses for LLM Agents.☆74Updated this week
- Dropbox LLM Security research code and results☆219Updated 7 months ago