xsankar / AI-Red-TeamingLinks
All things specific to LLM Red Teaming Generative AI
☆29Updated last year
Alternatives and similar repositories for AI-Red-Teaming
Users that are interested in AI-Red-Teaming are comparing it to the libraries listed below
Sorting:
- A productionized greedy coordinate gradient (GCG) attack tool for large language models (LLMs)☆152Updated last year
- Payloads for Attacking Large Language Models☆118Updated 2 weeks ago
- using ML models for red teaming☆45Updated 2 years ago
- ☆79Updated last month
- source code for the offsecml framework☆44Updated last year
- Reference notes for Attacking and Defending Generative AI presentation☆69Updated last year
- Example agents for the Dreadnode platform☆22Updated last month
- LLM | Security | Operations in one github repo with good links and pictures.☆86Updated last week
- ☆134Updated last week
- Code Repository for: AIRTBench: Measuring Autonomous AI Red Teaming Capabilities in Language Models☆92Updated this week
- ☆132Updated 6 months ago
- ☆44Updated last year
- Tree of Attacks (TAP) Jailbreaking Implementation☆117Updated last year
- A curated list of awesome AI Red Teaming resources and tools.☆29Updated 2 years ago
- https://arxiv.org/abs/2412.02776☆67Updated last year
- Prototype of Full Agentic Application Security Testing, FAAST = SAST + DAST + LLM agents☆67Updated 8 months ago
- Awesome products for securing AI systems includes open source and commercial options and an infographic licensed CC-BY-SA-4.0.☆82Updated last year
- Repository for CoSAI Workstream 4, Secure Design Patterns for Agentic Systems☆67Updated 2 weeks ago
- Data Scientists Go To Jupyter☆68Updated 10 months ago
- Delving into the Realm of LLM Security: An Exploration of Offensive and Defensive Tools, Unveiling Their Present Capabilities.☆166Updated 2 years ago
- ☆22Updated last year
- A LLM explicitly designed for getting hacked☆166Updated 2 years ago
- ☆20Updated 9 months ago
- General research for Dreadnode☆27Updated last year
- [IJCAI 2024] Imperio is an LLM-powered backdoor attack. It allows the adversary to issue language-guided instructions to control the vict…☆44Updated 11 months ago
- CALDERA plugin for adversary emulation of AI-enabled systems☆109Updated 2 years ago
- LLM Testing Findings Templates☆75Updated last year
- Secure Jupyter Notebooks and Experimentation Environment☆84Updated 11 months ago
- ATLAS tactics, techniques, and case studies data☆99Updated last month
- An interactive CLI application for interacting with authenticated Jupyter instances.☆55Updated 8 months ago