shlomihod / awesome-ai-red-teaming
A curated list of awesome AI Red Teaming resources and tools.
β10Updated last year
Related projects β
Alternatives and complementary repositories for awesome-ai-red-teaming
- π§ LLMFuzzer - Fuzzing Framework for Large Language Models π§ LLMFuzzer is the first open-source fuzzing framework specifically designed β¦β231Updated 9 months ago
- All things specific to LLM Red Teaming Generative AIβ14Updated 3 weeks ago
- Dropbox LLM Security research code and resultsβ216Updated 5 months ago
- A curated list of MLSecOps tools, articles and other resources on security applied to Machine Learning and MLOps systems.β240Updated last month
- Tree of Attacks (TAP) Jailbreaking Implementationβ95Updated 9 months ago
- β61Updated last month
- β‘ Vigil β‘ Detect prompt injections, jailbreaks, and other potentially risky Large Language Model (LLM) inputsβ309Updated 9 months ago
- LLM security and privacyβ40Updated 3 weeks ago
- future-proof vulnerability detection benchmark, based on CVEs in open-source reposβ44Updated this week
- β21Updated last month
- Every practical and proposed defense against prompt injection.β339Updated 5 months ago
- β178Updated 10 months ago
- This repository provides implementation to formalize and benchmark Prompt Injection attacks and defensesβ142Updated 2 months ago
- The automated prompt injection framework for LLM-integrated applications.β161Updated 2 months ago
- β57Updated 2 weeks ago
- XBOW Validation Benchmarksβ52Updated 2 months ago
- A repository of Language Model Vulnerabilities and Exposures (LVEs).β107Updated 8 months ago
- CTF challenges designed and implemented in machine learning applicationsβ111Updated 2 months ago
- Delving into the Realm of LLM Security: An Exploration of Offensive and Defensive Tools, Unveiling Their Present Capabilities.β149Updated last year
- Learn AI security through a series of vulnerable LLM CTF challenges. No sign ups, no cloud fees, run everything locally on your system.β247Updated 2 months ago
- β93Updated last month
- A benchmark for prompt injection detection systems.β86Updated 2 months ago
- β38Updated 4 months ago
- An overview of LLMs for cybersecurity.β412Updated last month
- This is a dataset intended to train a LLM model for a completely CVE focused input and output.β44Updated this week
- A PPO agent leveraging reinforcement learning performs Penetration Testing in a simulated computer network environment. The agent is traiβ¦β21Updated last year
- Adversarial Machine Learning (AML) Capture the Flag (CTF)β94Updated 8 months ago
- This project investigates the security of large language models by performing binary classification of a set of input prompts to discoverβ¦β34Updated 10 months ago
- Risks and targets for assessing LLMs & LLM vulnerabilitiesβ25Updated 5 months ago
- A collection of awesome resources related AI securityβ124Updated 7 months ago