humane-intelligence / ai_village_defcon_grt_dataLinks
☆14Updated last year
Alternatives and similar repositories for ai_village_defcon_grt_data
Users that are interested in ai_village_defcon_grt_data are comparing it to the libraries listed below
Sorting:
- General research for Dreadnode☆27Updated last year
- ☆66Updated 4 months ago
- All things specific to LLM Red Teaming Generative AI☆29Updated last year
- ☆132Updated 6 months ago
- Example agents for the Dreadnode platform☆22Updated last month
- Tree of Attacks (TAP) Jailbreaking Implementation☆117Updated last year
- CyberGym is a large-scale, high-quality cybersecurity evaluation framework designed to rigorously assess the capabilities of AI agents on…☆106Updated 2 weeks ago
- Implementation of BEAST adversarial attack for language models (ICML 2024)☆92Updated last year
- A collection of prompt injection mitigation techniques.☆26Updated 2 years ago
- Code for the paper "Defeating Prompt Injections by Design"☆220Updated 7 months ago
- A repository of Language Model Vulnerabilities and Exposures (LVEs).☆112Updated last year
- Codebase of https://arxiv.org/abs/2410.14923☆54Updated last year
- source code for the offsecml framework☆44Updated last year
- Code snippets to reproduce MCP tool poisoning attacks.☆192Updated 9 months ago
- Data Scientists Go To Jupyter☆68Updated 10 months ago
- A productionized greedy coordinate gradient (GCG) attack tool for large language models (LLMs)☆155Updated last year
- A YAML based format for describing tools to LLMs, like man pages but for robots!☆83Updated 9 months ago
- ☆22Updated last year
- A benchmark for prompt injection detection systems.☆156Updated last month
- A Dynamic Environment to Evaluate Attacks and Defenses for LLM Agents.☆420Updated 2 months ago
- Repository for CoSAI Workstream 4, Secure Design Patterns for Agentic Systems☆82Updated 2 weeks ago
- https://arxiv.org/abs/2412.02776☆67Updated last year
- A benchmark for evaluating the robustness of LLMs and defenses to indirect prompt injection attacks.☆103Updated last year
- Code Repository for: AIRTBench: Measuring Autonomous AI Red Teaming Capabilities in Language Models☆92Updated this week
- LobotoMl is a set of scripts and tools to assess production deployments of ML services☆10Updated 3 years ago
- Payloads for Attacking Large Language Models☆118Updated 2 weeks ago
- ☆137Updated this week
- The automated prompt injection framework for LLM-integrated applications.☆251Updated last year
- ☆82Updated last month
- future-proof vulnerability detection benchmark, based on CVEs in open-source repos☆64Updated last week