microsoft / gandalf_vs_gandalfLinks
Turning Gandalf against itself. Use LLMs to automate playing Lakera Gandalf challenge without needing to set up an account with a platform provider.
☆28Updated last year
Alternatives and similar repositories for gandalf_vs_gandalf
Users that are interested in gandalf_vs_gandalf are comparing it to the libraries listed below
Sorting:
- ☆52Updated 2 months ago
- HoneyAgents is a PoC demo of an AI-driven system that combines honeypots with autonomous AI agents to detect and mitigate cyber threats. …☆53Updated last year
- Make your GenAI Apps Safe & Secure Test & harden your system prompt☆519Updated last month
- Dropbox LLM Security research code and results☆228Updated last year
- Test Software for the Characterization of AI Technologies☆260Updated this week
- [Corca / ML] Automatically solved Gandalf AI with LLM☆50Updated 2 years ago
- Red-Teaming Language Models with DSPy☆202Updated 5 months ago
- ⚡ Vigil ⚡ Detect prompt injections, jailbreaks, and other potentially risky Large Language Model (LLM) inputs☆396Updated last year
- A benchmark for prompt injection detection systems.☆122Updated 2 months ago
- ComPromptMized: Unleashing Zero-click Worms that Target GenAI-Powered Applications☆203Updated last year
- Top 10 for Agentic AI (AI Agent Security) serves as the core for OWASP and CSA Red teaming work☆119Updated last month
- Risks and targets for assessing LLMs & LLM vulnerabilities☆31Updated last year
- Curated list of Open Source project focused on LLM security☆50Updated 8 months ago
- The fastest Trust Layer for AI Agents☆138Updated last month
- Moonshot - A simple and modular tool to evaluate and red-team any LLM application.☆258Updated this week
- Project Mantis: Hacking Back the AI-Hacker; Prompt Injection as a Defense Against LLM-driven Cyberattacks☆71Updated last month
- The project serves as a strategic advisory tool, capitalizing on the ZySec series of AI models to amplify the capabilities of security pr…☆53Updated last year
- A repository of Language Model Vulnerabilities and Exposures (LVEs).☆112Updated last year
- OWASP Foundation Web Respository☆284Updated this week
- a prompt injection scanner for custom LLM applications☆835Updated 4 months ago
- ☆254Updated 6 months ago
- ATLAS tactics, techniques, and case studies data☆77Updated 2 months ago
- Every practical and proposed defense against prompt injection.☆495Updated 4 months ago
- Prompt Injection Primer for Engineers☆443Updated last year
- A toolkit for detecting and protecting against vulnerabilities in Large Language Models (LLMs).☆139Updated last year
- Explore AI Supply Chain Risk with the AI Risk Database☆59Updated last year
- Delving into the Realm of LLM Security: An Exploration of Offensive and Defensive Tools, Unveiling Their Present Capabilities.☆163Updated last year
- Lightweight LLM Interaction Framework☆296Updated this week
- 📚 A curated list of papers & technical articles on AI Quality & Safety☆188Updated 3 months ago
- 🤖 A GitHub action that leverages fabric patterns through an agent-based approach☆28Updated 6 months ago