humane-intelligence / ai_village_defcon_grt_data
☆13Updated 10 months ago
Alternatives and similar repositories for ai_village_defcon_grt_data:
Users that are interested in ai_village_defcon_grt_data are comparing it to the libraries listed below
- General research for Dreadnode☆21Updated 10 months ago
- ☆64Updated 3 months ago
- Tree of Attacks (TAP) Jailbreaking Implementation☆106Updated last year
- source code for the offsecml framework☆38Updated 10 months ago
- ☆13Updated 4 months ago
- https://arxiv.org/abs/2412.02776☆52Updated 4 months ago
- Data Scientists Go To Jupyter☆62Updated last month
- A productionized greedy coordinate gradient (GCG) attack tool for large language models (LLMs)☆98Updated 4 months ago
- A repository of Language Model Vulnerabilities and Exposures (LVEs).☆109Updated last year
- ☆34Updated 2 months ago
- Integrate PyRIT in existing tools☆22Updated last month
- ☆16Updated 10 months ago
- future-proof vulnerability detection benchmark, based on CVEs in open-source repos☆52Updated last week
- All things specific to LLM Red Teaming Generative AI☆24Updated 6 months ago
- A benchmark for prompt injection detection systems.☆100Updated 2 months ago
- The official implementation of our pre-print paper "Automatic and Universal Prompt Injection Attacks against Large Language Models".☆45Updated 6 months ago
- The D-CIPHER and NYU CTF baseline LLM Agents built for NYU CTF Bench☆65Updated 2 weeks ago
- A collection of prompt injection mitigation techniques.☆22Updated last year
- ☆93Updated last month
- A utility to inspect, validate, sign and verify machine learning model files.☆56Updated 2 months ago
- ☆37Updated 6 months ago
- [NDSS'25 Best Technical Poster] A collection of automated evaluators for assessing jailbreak attempts.☆148Updated 3 weeks ago
- A YAML based format for describing tools to LLMs, like man pages but for robots!☆69Updated 2 weeks ago
- ChainReactor is a research project that leverages AI planning to discover exploitation chains for privilege escalation on Unix systems. T…☆44Updated 5 months ago
- ☆33Updated 6 months ago
- Implementation of BEAST adversarial attack for language models (ICML 2024)☆82Updated 11 months ago
- Codebase of https://arxiv.org/abs/2410.14923☆46Updated 6 months ago
- LLM security and privacy☆48Updated 6 months ago
- The automated prompt injection framework for LLM-integrated applications.☆198Updated 7 months ago
- ☆59Updated 5 months ago