amazon-science / TurboFuzzLLMLinks
TurboFuzzLLM: Turbocharging Mutation-based Fuzzing for Effectively Jailbreaking Large Language Models in Practice
☆20Updated 3 weeks ago
Alternatives and similar repositories for TurboFuzzLLM
Users that are interested in TurboFuzzLLM are comparing it to the libraries listed below
Sorting:
- Implementation of BEAST adversarial attack for language models (ICML 2024)☆92Updated last year
- A repository of Language Model Vulnerabilities and Exposures (LVEs).☆112Updated last year
- ☆179Updated 6 months ago
- General research for Dreadnode☆27Updated last year
- PAL: Proxy-Guided Black-Box Attack on Large Language Models☆55Updated last year
- Code for the paper "Defeating Prompt Injections by Design"☆179Updated 5 months ago
- ☆98Updated 4 months ago
- TaskTracker is an approach to detecting task drift in Large Language Models (LLMs) by analysing their internal activations. It provides a…☆78Updated 3 months ago
- ☆35Updated last year
- [ICML 2024] COLD-Attack: Jailbreaking LLMs with Stealthiness and Controllability☆173Updated last year
- ☆77Updated last month
- LLM security and privacy☆52Updated last year
- A Dynamic Environment to Evaluate Attacks and Defenses for LLM Agents.☆379Updated 2 weeks ago
- Papers about red teaming LLMs and Multimodal models.☆157Updated 6 months ago
- Tree of Attacks (TAP) Jailbreaking Implementation☆116Updated last year
- Official repository for the paper "ALERT: A Comprehensive Benchmark for Assessing Large Language Models’ Safety through Red Teaming"☆51Updated last year
- The official implementation of our pre-print paper "Automatic and Universal Prompt Injection Attacks against Large Language Models".☆67Updated last year
- A benchmark for evaluating the robustness of LLMs and defenses to indirect prompt injection attacks.☆91Updated last year
- Code snippets to reproduce MCP tool poisoning attacks.☆187Updated 8 months ago
- Official implementation of the WASP web agent security benchmark☆60Updated 4 months ago
- WMDP is a LLM proxy benchmark for hazardous knowledge in bio, cyber, and chemical security. We also release code for RMU, an unlearning m…☆156Updated 6 months ago
- A re-implementation of the "Red Teaming Language Models with Language Models" paper by Perez et al., 2022☆35Updated 2 years ago
- PhD/MSc course on Machine Learning Security (Univ. Cagliari)☆221Updated last week
- ☆21Updated last year
- This repository provides a benchmark for prompt injection attacks and defenses in LLMs☆361Updated last month
- SECURE: Benchmarking Generative Large Language Models as a Cyber Advisory☆15Updated last year
- CyberGym is a large-scale, high-quality cybersecurity evaluation framework designed to rigorously assess the capabilities of AI agents on…☆100Updated 2 months ago
- Risks and targets for assessing LLMs & LLM vulnerabilities☆33Updated last year
- Cyber-Zero: Training Cybersecurity Agents Without Runtime☆48Updated 2 weeks ago
- A benchmark for prompt injection detection systems.☆151Updated 3 months ago