xsankar / AI-Red-Teaming
All things specific to LLM Red Teaming Generative AI
β21Updated 3 months ago
Alternatives and similar repositories for AI-Red-Teaming:
Users that are interested in AI-Red-Teaming are comparing it to the libraries listed below
- π€π‘οΈπππ Tiny package designed to support red teams and penetration testers in exploiting large language model AI solutions.β22Updated 9 months ago
- The D-CIPHER and NYU CTF baseline LLM Agents built for NYU CTF Benchβ49Updated 2 weeks ago
- Payloads for Attacking Large Language Modelsβ74Updated 7 months ago
- The automated prompt injection framework for LLM-integrated applications.β185Updated 5 months ago
- A collection of awesome resources related AI securityβ174Updated 2 weeks ago
- π§ LLMFuzzer - Fuzzing Framework for Large Language Models π§ LLMFuzzer is the first open-source fuzzing framework specifically designed β¦β258Updated last year
- A productionized greedy coordinate gradient (GCG) attack tool for large language models (LLMs)β84Updated 2 months ago
- CTF challenges designed and implemented in machine learning applicationsβ131Updated 5 months ago
- A curated list of awesome resources about LLM supply chain security (including papers, security reports and CVEs)β38Updated last month
- A collection of prompt injection mitigation techniques.β20Updated last year
- using ML models for red teamingβ42Updated last year
- A collection of agents that use Large Language Models (LLMs) to perform tasks common on our day to day jobs in cyber security.β85Updated 9 months ago
- [USENIX Security '24] An LLM-Assisted Easy-to-Trigger Backdoor Attack on Code Completion Models: Injecting Disguised Vulnerabilities agaiβ¦β36Updated 3 months ago
- Tree of Attacks (TAP) Jailbreaking Implementationβ99Updated last year
- A benchmark for prompt injection detection systems.β96Updated 2 weeks ago
- A LLM explicitly designed for getting hackedβ136Updated last year
- Data Scientists Go To Jupyterβ62Updated 2 months ago
- This repository provides implementation to formalize and benchmark Prompt Injection attacks and defensesβ172Updated 3 weeks ago
- β13Updated 8 months ago
- General research for Dreadnodeβ19Updated 8 months ago
- [IJCAI 2024] Imperio is an LLM-powered backdoor attack. It allows the adversary to issue language-guided instructions to control the victβ¦β41Updated this week
- β64Updated last month
- β34Updated 2 weeks ago
- [CCS'24] An LLM-based, fully automated fuzzing tool for option combination testing.β64Updated last month
- LLM | Security | Operations in one github repo with good links and pictures.β24Updated last month
- XBOW Validation Benchmarksβ71Updated 5 months ago
- VulZoo: A Comprehensive Vulnerability Intelligence Dataset (ASE 2024 Demo)β30Updated 3 months ago
- source code for the offsecml frameworkβ37Updated 8 months ago
- β118Updated 3 months ago
- A repository of Language Model Vulnerabilities and Exposures (LVEs).β108Updated 11 months ago