humane-intelligence / ai_village_defcon_grt_data
☆13Updated 5 months ago
Related projects ⓘ
Alternatives and complementary repositories for ai_village_defcon_grt_data
- ☆62Updated last month
- General research for Dreadnode☆17Updated 5 months ago
- A repository of Language Model Vulnerabilities and Exposures (LVEs).☆107Updated 8 months ago
- Implementation of BEAST adversarial attack for language models (ICML 2024)☆73Updated 6 months ago
- Tree of Attacks (TAP) Jailbreaking Implementation☆95Updated 9 months ago
- ☆36Updated this week
- A benchmark for prompt injection detection systems.☆87Updated 2 months ago
- Payloads for Attacking Large Language Models☆64Updated 4 months ago
- source code for the offsecml framework☆35Updated 5 months ago
- Data Scientists Go To Jupyter☆57Updated last week
- ATLAS tactics, techniques, and case studies data☆49Updated last month
- future-proof vulnerability detection benchmark, based on CVEs in open-source repos☆44Updated last week
- XBOW Validation Benchmarks☆53Updated 2 months ago
- ☆22Updated last month
- ☆63Updated this week
- [USENIX Security '24] An LLM-Assisted Easy-to-Trigger Backdoor Attack on Code Completion Models: Injecting Disguised Vulnerabilities agai…☆25Updated 3 weeks ago
- CTF challenges designed and implemented in machine learning applications☆114Updated 2 months ago
- LLM security and privacy☆41Updated last month
- A collection of prompt injection mitigation techniques.☆18Updated last year
- The automated prompt injection framework for LLM-integrated applications.☆163Updated 2 months ago
- The official implementation of our pre-print paper "Automatic and Universal Prompt Injection Attacks against Large Language Models".☆34Updated 3 weeks ago
- Risks and targets for assessing LLMs & LLM vulnerabilities☆25Updated 5 months ago
- A utility to inspect, validate, sign and verify machine learning model files.☆42Updated 2 weeks ago
- ☆38Updated 4 months ago
- TAP: An automated jailbreaking method for black-box LLMs☆119Updated 8 months ago
- ☆20Updated 2 months ago
- A Dynamic Environment to Evaluate Attacks and Defenses for LLM Agents.☆65Updated this week
- This repository provides implementation to formalize and benchmark Prompt Injection attacks and defenses☆146Updated 2 months ago
- ☆26Updated this week
- A benchmark for evaluating the robustness of LLMs and defenses to indirect prompt injection attacks.☆47Updated 7 months ago