microsoft / gandalf_vs_gandalf
Turning Gandalf against itself. Use LLMs to automate playing Lakera Gandalf challenge without needing to set up an account with a platform provider.
☆29Updated last year
Alternatives and similar repositories for gandalf_vs_gandalf:
Users that are interested in gandalf_vs_gandalf are comparing it to the libraries listed below
- HoneyAgents is a PoC demo of an AI-driven system that combines honeypots with autonomous AI agents to detect and mitigate cyber threats. …☆42Updated last year
- [Corca / ML] Automatically solved Gandalf AI with LLM☆48Updated last year
- Red-Teaming Language Models with DSPy☆175Updated last month
- Dropbox LLM Security research code and results☆221Updated 10 months ago
- A benchmark for prompt injection detection systems.☆99Updated last month
- ☆37Updated 3 months ago
- ☆30Updated last month
- Project LLM Verification Standard☆41Updated 11 months ago
- Make your GenAI Apps Safe & Secure Test & harden your system prompt☆453Updated 5 months ago
- ☆37Updated 4 months ago
- OWASP Machine Learning Security Top 10 Project☆83Updated 2 months ago
- Payloads for Attacking Large Language Models☆77Updated 8 months ago
- Test Software for the Characterization of AI Technologies☆243Updated this week
- ComPromptMized: Unleashing Zero-click Worms that Target GenAI-Powered Applications☆199Updated last year
- A toolkit for detecting and protecting against vulnerabilities in Large Language Models (LLMs).☆135Updated last year
- A powerful tool that leverages AI to automatically generate comprehensive security documentation for your projects☆62Updated 2 weeks ago
- Risks and targets for assessing LLMs & LLM vulnerabilities☆30Updated 10 months ago
- Project Mantis: Hacking Back the AI-Hacker; Prompt Injection as a Defense Against LLM-driven Cyberattacks☆66Updated 3 months ago
- A collection of awesome resources related AI security☆195Updated last month
- The project serves as a strategic advisory tool, capitalizing on the ZySec series of AI models to amplify the capabilities of security pr…☆45Updated 10 months ago
- Curation of prompts that are known to be adversarial to large language models☆179Updated 2 years ago
- Top 10 for Agentic AI (AI Agent Security) - Pre-release version☆72Updated 3 weeks ago
- This project investigates the security of large language models by performing binary classification of a set of input prompts to discover…☆38Updated last year
- Detect AI-generated text [relatively] quickly via compression ratios☆232Updated last month
- Masked Python SDK wrapper for OpenAI API. Use public LLM APIs securely.☆116Updated 2 years ago
- Prompt Injection Primer for Engineers☆423Updated last year
- A writeup for the Gandalf prompt injection game.☆37Updated last year
- Guard your LangChain applications against prompt injection with Lakera ChainGuard.☆20Updated 3 weeks ago
- Explore AI Supply Chain Risk with the AI Risk Database☆53Updated 10 months ago
- A guide to LLM hacking: fundamentals, prompt injection, offense, and defense☆148Updated last year