safellama / plexiglass
A toolkit for detecting and protecting against vulnerabilities in Large Language Models (LLMs).
☆125Updated last year
Alternatives and similar repositories for plexiglass:
Users that are interested in plexiglass are comparing it to the libraries listed below
- Red-Teaming Language Models with DSPy☆153Updated 9 months ago
- Risks and targets for assessing LLMs & LLM vulnerabilities☆30Updated 7 months ago
- Lightweight LLM Interaction Framework☆225Updated this week
- Fiddler Auditor is a tool to evaluate language models.☆174Updated 10 months ago
- ⚡ Vigil ⚡ Detect prompt injections, jailbreaks, and other potentially risky Large Language Model (LLM) inputs☆339Updated 11 months ago
- A repository of Language Model Vulnerabilities and Exposures (LVEs).☆108Updated 10 months ago
- Machine Learning Attack Series☆57Updated 8 months ago
- ☆26Updated 2 months ago
- PromptInject is a framework that assembles prompts in a modular fashion to provide a quantitative analysis of the robustness of LLMs to a…☆323Updated 10 months ago
- Curation of prompts that are known to be adversarial to large language models☆177Updated last year
- ComPromptMized: Unleashing Zero-click Worms that Target GenAI-Powered Applications☆197Updated 10 months ago
- Helps you build better AI agents through debuggable unit testing☆141Updated this week
- The fastest && easiest LLM security guardrails for CX AI Agents and applications.☆114Updated this week
- source for llmsec.net☆13Updated 5 months ago
- TaskTracker is an approach to detecting task drift in Large Language Models (LLMs) by analysing their internal activations. It provides a…☆42Updated last month
- Dropbox LLM Security research code and results☆219Updated 7 months ago
- HoneyAgents is a PoC demo of an AI-driven system that combines honeypots with autonomous AI agents to detect and mitigate cyber threats. …☆39Updated last year
- Turning Gandalf against itself. Use LLMs to automate playing Lakera Gandalf challenge without needing to set up an account with a platfor…☆27Updated last year
- Make your GenAI Apps Safe & Secure Test & harden your system prompt☆429Updated 3 months ago
- A subset of jailbreaks automatically discovered by the Haize Labs haizing suite.☆88Updated 7 months ago
- LLM security and privacy☆43Updated 3 months ago
- A powerful tool for automated LLM fuzzing. It is designed to help developers and security researchers identify and mitigate potential jai…☆230Updated this week
- A benchmark for prompt injection detection systems.☆94Updated 4 months ago
- ☆67Updated last month
- An open-source compliance-centered evaluation framework for Generative AI models☆121Updated last month
- Masked Python SDK wrapper for OpenAI API. Use public LLM APIs securely.☆115Updated last year
- Test Software for the Characterization of AI Technologies☆235Updated this week
- Whispers in the Machine: Confidentiality in LLM-integrated Systems☆31Updated last month
- Practical examples of "Flawed Machine Learning Security" together with ML Security best practice across the end to end stages of the mach…☆104Updated 2 years ago
- A prompt injection game to collect data for robust ML research☆49Updated 3 weeks ago