microsoft / gandalf_vs_gandalf
Turning Gandalf against itself. Use LLMs to automate playing Lakera Gandalf challenge without needing to set up an account with a platform provider.
☆29Updated last year
Alternatives and similar repositories for gandalf_vs_gandalf:
Users that are interested in gandalf_vs_gandalf are comparing it to the libraries listed below
- [Corca / ML] Automatically solved Gandalf AI with LLM☆48Updated last year
- HoneyAgents is a PoC demo of an AI-driven system that combines honeypots with autonomous AI agents to detect and mitigate cyber threats. …☆41Updated last year
- Guard your LangChain applications against prompt injection with Lakera ChainGuard.☆18Updated last month
- ComPromptMized: Unleashing Zero-click Worms that Target GenAI-Powered Applications☆197Updated 11 months ago
- ☆27Updated 3 months ago
- A benchmark for prompt injection detection systems.☆96Updated 2 weeks ago
- Red-Teaming Language Models with DSPy☆169Updated last week
- ☆36Updated 2 months ago
- Every practical and proposed defense against prompt injection.☆389Updated 8 months ago
- Payloads for Attacking Large Language Models☆74Updated 7 months ago
- Dropbox LLM Security research code and results☆220Updated 9 months ago
- source for llmsec.net☆14Updated 6 months ago
- OWASP Top 10 for Agentic AI (AI Agent Security) - Pre-release version☆53Updated this week
- Lakera - ChatGPT Data Leak Protection☆22Updated 7 months ago
- ATLAS tactics, techniques, and case studies data☆56Updated 4 months ago
- Make your GenAI Apps Safe & Secure Test & harden your system prompt☆435Updated 4 months ago
- ⚡ Vigil ⚡ Detect prompt injections, jailbreaks, and other potentially risky Large Language Model (LLM) inputs☆350Updated last year
- Initiative to evaluate and rank the most popular LLMs across common task types based on their propensity to hallucinate.☆106Updated 5 months ago
- Scripts and Content for working with Open AI☆154Updated 4 months ago
- Approximation of the Claude 3 tokenizer by inspecting generation stream☆123Updated 7 months ago
- A toolkit for detecting and protecting against vulnerabilities in Large Language Models (LLMs).☆130Updated last year
- Curation of prompts that are known to be adversarial to large language models☆179Updated 2 years ago
- A writeup for the Gandalf prompt injection game.☆36Updated last year
- Learn about a type of vulnerability that specifically targets machine learning models☆220Updated 8 months ago
- Project LLM Verification Standard☆38Updated 10 months ago
- ☆28Updated 3 months ago
- ☆70Updated 4 months ago
- Awesome products for securing AI systems includes open source and commercial options and an infographic licensed CC-BY-SA-4.0.☆58Updated 8 months ago
- The project serves as a strategic advisory tool, capitalizing on the ZySec series of AI models to amplify the capabilities of security pr…☆43Updated 9 months ago
- A subset of jailbreaks automatically discovered by the Haize Labs haizing suite.☆89Updated 8 months ago