safellama / plexiglass
A toolkit for detecting and protecting against vulnerabilities in Large Language Models (LLMs).
☆136Updated last year
Alternatives and similar repositories for plexiglass:
Users that are interested in plexiglass are comparing it to the libraries listed below
- Red-Teaming Language Models with DSPy☆183Updated 2 months ago
- Risks and targets for assessing LLMs & LLM vulnerabilities☆30Updated 10 months ago
- The fastest Trust Layer for AI Agents☆130Updated last month
- ATLAS tactics, techniques, and case studies data☆63Updated last month
- ComPromptMized: Unleashing Zero-click Worms that Target GenAI-Powered Applications☆201Updated last year
- A benchmark for prompt injection detection systems.☆100Updated 2 months ago
- ⚡ Vigil ⚡ Detect prompt injections, jailbreaks, and other potentially risky Large Language Model (LLM) inputs☆378Updated last year
- A collection of prompt injection mitigation techniques.☆22Updated last year
- source for llmsec.net☆15Updated 9 months ago
- Fiddler Auditor is a tool to evaluate language models.☆179Updated last year
- A JupyterLab extension to evaluate the security of your Jupyter environment☆39Updated last year
- Test Software for the Characterization of AI Technologies☆246Updated last week
- A repository of Language Model Vulnerabilities and Exposures (LVEs).☆109Updated last year
- ☆31Updated 5 months ago
- HoneyAgents is a PoC demo of an AI-driven system that combines honeypots with autonomous AI agents to detect and mitigate cyber threats. …☆46Updated last year
- Dropbox LLM Security research code and results☆222Updated 11 months ago
- [Corca / ML] Automatically solved Gandalf AI with LLM☆49Updated last year
- A guide to LLM hacking: fundamentals, prompt injection, offense, and defense☆148Updated 2 years ago
- Lightweight LLM Interaction Framework☆262Updated this week
- Project Mantis: Hacking Back the AI-Hacker; Prompt Injection as a Defense Against LLM-driven Cyberattacks☆67Updated 4 months ago
- Payloads for Attacking Large Language Models☆79Updated 9 months ago
- Codebase of https://arxiv.org/abs/2410.14923☆46Updated 6 months ago
- Explore AI Supply Chain Risk with the AI Risk Database☆53Updated 11 months ago
- Machine Learning Attack Series☆60Updated 11 months ago
- PromptInject is a framework that assembles prompts in a modular fashion to provide a quantitative analysis of the robustness of LLMs to a…☆361Updated last year
- Tree of Attacks (TAP) Jailbreaking Implementation☆106Updated last year
- An open-source compliance-centered evaluation framework for Generative AI models☆147Updated 4 months ago
- OWASP Machine Learning Security Top 10 Project☆83Updated 2 months ago
- LLM security and privacy☆48Updated 6 months ago
- A collection of agents that use Large Language Models (LLMs) to perform tasks common on our day to day jobs in cyber security.☆103Updated 11 months ago