safellama / plexiglassLinks
A toolkit for detecting and protecting against vulnerabilities in Large Language Models (LLMs).
☆152Updated this week
Alternatives and similar repositories for plexiglass
Users that are interested in plexiglass are comparing it to the libraries listed below
Sorting:
- A repository of Language Model Vulnerabilities and Exposures (LVEs).☆112Updated last year
- Red-Teaming Language Models with DSPy☆250Updated 11 months ago
- The fastest Trust Layer for AI Agents☆152Updated this week
- Here Comes the AI Worm: Preventing the Propagation of Adversarial Self-Replicating Prompts Within GenAI Ecosystems☆222Updated 5 months ago
- Lightweight LLM Interaction Framework☆406Updated this week
- Risks and targets for assessing LLMs & LLM vulnerabilities☆33Updated last year
- Test Software for the Characterization of AI Technologies☆277Updated this week
- A JupyterLab extension to evaluate the security of your Jupyter environment☆38Updated 2 years ago
- This project investigates the security of large language models by performing binary classification of a set of input prompts to discover…☆57Updated 2 years ago
- ⚡ Vigil ⚡ Detect prompt injections, jailbreaks, and other potentially risky Large Language Model (LLM) inputs☆452Updated 2 years ago
- Fiddler Auditor is a tool to evaluate language models.☆188Updated last year
- A guide to LLM hacking: fundamentals, prompt injection, offense, and defense☆182Updated 2 years ago
- Curation of prompts that are known to be adversarial to large language models☆188Updated 2 years ago
- Dropbox LLM Security research code and results☆254Updated last year
- A collection of prompt injection mitigation techniques.☆27Updated 2 years ago
- source for llmsec.net☆16Updated last year
- A benchmark for prompt injection detection systems.☆158Updated last month
- ☆54Updated last week
- BlindBox is a tool to isolate and deploy applications inside Trusted Execution Environments for privacy-by-design apps☆64Updated 2 years ago
- Codebase of https://arxiv.org/abs/2410.14923☆54Updated last year
- LLM security and privacy☆54Updated last year
- 📚 A curated list of papers & technical articles on AI Quality & Safety☆200Updated 9 months ago
- PromptInject is a framework that assembles prompts in a modular fashion to provide a quantitative analysis of the robustness of LLMs to a…☆452Updated last year
- ATLAS tactics, techniques, and case studies data☆99Updated last week
- HoneyAgents is a PoC demo of an AI-driven system that combines honeypots with autonomous AI agents to detect and mitigate cyber threats. …☆59Updated 2 years ago
- Code for the paper "Defeating Prompt Injections by Design"☆246Updated 7 months ago
- Explore AI Supply Chain Risk with the AI Risk Database☆65Updated last year
- A subset of jailbreaks automatically discovered by the Haize Labs haizing suite.☆100Updated 9 months ago
- Machine Learning Attack Series☆75Updated last year
- This repository contains various attack against Large Language Models.☆128Updated last year