humane-intelligence / ai_village_defcon_grt_dataLinks
☆13Updated 11 months ago
Alternatives and similar repositories for ai_village_defcon_grt_data
Users that are interested in ai_village_defcon_grt_data are comparing it to the libraries listed below
Sorting:
- General research for Dreadnode☆23Updated 11 months ago
- All things specific to LLM Red Teaming Generative AI☆25Updated 7 months ago
- Implementation of BEAST adversarial attack for language models (ICML 2024)☆87Updated last year
- Papers about red teaming LLMs and Multimodal models.☆121Updated last week
- Tree of Attacks (TAP) Jailbreaking Implementation☆109Updated last year
- TaskTracker is an approach to detecting task drift in Large Language Models (LLMs) by analysing their internal activations. It provides a…☆56Updated 2 months ago
- source code for the offsecml framework☆40Updated last year
- ☆65Updated 4 months ago
- ☆40Updated 8 months ago
- A repository of Language Model Vulnerabilities and Exposures (LVEs).☆110Updated last year
- A benchmark for evaluating the robustness of LLMs and defenses to indirect prompt injection attacks.☆69Updated last year
- The D-CIPHER and NYU CTF baseline LLM Agents built for NYU CTF Bench☆77Updated last month
- A benchmark for prompt injection detection systems.☆115Updated 3 weeks ago
- A productionized greedy coordinate gradient (GCG) attack tool for large language models (LLMs)☆113Updated 5 months ago
- ☆63Updated 11 months ago
- ☆109Updated 2 weeks ago
- The official implementation of our pre-print paper "Automatic and Universal Prompt Injection Attacks against Large Language Models".☆49Updated 7 months ago
- An interactive CLI application for interacting with authenticated Jupyter instances.☆53Updated last month
- ☆48Updated last week
- future-proof vulnerability detection benchmark, based on CVEs in open-source repos☆56Updated last week
- ATLAS tactics, techniques, and case studies data☆73Updated last month
- Payloads for Attacking Large Language Models☆89Updated 10 months ago
- ☆44Updated 10 months ago
- Data Scientists Go To Jupyter☆64Updated 3 months ago
- A collection of prompt injection mitigation techniques.☆23Updated last year
- PAL: Proxy-Guided Black-Box Attack on Large Language Models☆51Updated 9 months ago
- https://arxiv.org/abs/2412.02776☆54Updated 6 months ago
- A Dynamic Environment to Evaluate Attacks and Defenses for LLM Agents.☆175Updated this week
- TAP: An automated jailbreaking method for black-box LLMs☆171Updated 5 months ago
- Risks and targets for assessing LLMs & LLM vulnerabilities☆30Updated last year