microsoft / gandalf_vs_gandalfLinks
Turning Gandalf against itself. Use LLMs to automate playing Lakera Gandalf challenge without needing to set up an account with a platform provider.
☆28Updated last year
Alternatives and similar repositories for gandalf_vs_gandalf
Users that are interested in gandalf_vs_gandalf are comparing it to the libraries listed below
Sorting:
- ☆54Updated 4 months ago
- HoneyAgents is a PoC demo of an AI-driven system that combines honeypots with autonomous AI agents to detect and mitigate cyber threats. …☆55Updated last year
- [Corca / ML] Automatically solved Gandalf AI with LLM☆51Updated 2 years ago
- Make your GenAI Apps Safe & Secure Test & harden your system prompt☆553Updated last month
- Dropbox LLM Security research code and results☆233Updated last year
- A benchmark for prompt injection detection systems.☆128Updated last week
- Red-Teaming Language Models with DSPy☆211Updated 6 months ago
- ⚡ Vigil ⚡ Detect prompt injections, jailbreaks, and other potentially risky Large Language Model (LLM) inputs☆410Updated last year
- OWASP Foundation Web Respository☆303Updated 2 weeks ago
- Here Comes the AI Worm: Preventing the Propagation of Adversarial Self-Replicating Prompts Within GenAI Ecosystems☆205Updated this week
- ☆38Updated 8 months ago
- A powerful tool that leverages AI to automatically generate comprehensive security documentation for your projects☆90Updated this week
- Test Software for the Characterization of AI Technologies☆261Updated last week
- Top 10 for Agentic AI (AI Agent Security) serves as the core for OWASP and CSA Red teaming work☆131Updated 2 months ago
- Curated list of Open Source project focused on LLM security☆62Updated 10 months ago
- ATLAS tactics, techniques, and case studies data☆77Updated 3 weeks ago
- Project LLM Verification Standard☆48Updated 3 months ago
- The project serves as a strategic advisory tool, capitalizing on the ZySec series of AI models to amplify the capabilities of security pr…☆58Updated last year
- Codebase of https://arxiv.org/abs/2410.14923☆50Updated 10 months ago
- The fastest Trust Layer for AI Agents☆142Updated 3 months ago
- Moonshot - A simple and modular tool to evaluate and red-team any LLM application.☆266Updated last month
- Every practical and proposed defense against prompt injection.☆537Updated 6 months ago
- A toolkit for detecting and protecting against vulnerabilities in Large Language Models (LLMs).☆142Updated last year
- Lightweight LLM Interaction Framework☆371Updated this week
- A guide to LLM hacking: fundamentals, prompt injection, offense, and defense☆168Updated 2 years ago
- Code Repository for: AIRTBench: Measuring Autonomous AI Red Teaming Capabilities in Language Models☆75Updated this week
- Reference notes for Attacking and Defending Generative AI presentation☆65Updated last year
- Project Mantis: Hacking Back the AI-Hacker; Prompt Injection as a Defense Against LLM-driven Cyberattacks☆79Updated 3 months ago
- ☆66Updated this week
- Zero Trust Agent☆28Updated this week