safellama / plexiglassLinks
A toolkit for detecting and protecting against vulnerabilities in Large Language Models (LLMs).
☆152Updated 2 years ago
Alternatives and similar repositories for plexiglass
Users that are interested in plexiglass are comparing it to the libraries listed below
Sorting:
- The fastest Trust Layer for AI Agents☆149Updated 8 months ago
- Red-Teaming Language Models with DSPy☆250Updated 11 months ago
- A repository of Language Model Vulnerabilities and Exposures (LVEs).☆112Updated last year
- Test Software for the Characterization of AI Technologies☆276Updated this week
- Lightweight LLM Interaction Framework☆406Updated this week
- A JupyterLab extension to evaluate the security of your Jupyter environment☆38Updated 2 years ago
- Here Comes the AI Worm: Preventing the Propagation of Adversarial Self-Replicating Prompts Within GenAI Ecosystems☆222Updated 4 months ago
- Dropbox LLM Security research code and results☆254Updated last year
- Risks and targets for assessing LLMs & LLM vulnerabilities☆33Updated last year
- ⚡ Vigil ⚡ Detect prompt injections, jailbreaks, and other potentially risky Large Language Model (LLM) inputs☆439Updated 2 years ago
- A collection of prompt injection mitigation techniques.☆26Updated 2 years ago
- Fiddler Auditor is a tool to evaluate language models.☆188Updated last year
- A benchmark for prompt injection detection systems.☆156Updated last month
- BlindBox is a tool to isolate and deploy applications inside Trusted Execution Environments for privacy-by-design apps☆64Updated 2 years ago
- Payloads for Attacking Large Language Models☆118Updated 2 weeks ago
- 📚 A curated list of papers & technical articles on AI Quality & Safety☆199Updated 9 months ago
- Curation of prompts that are known to be adversarial to large language models☆188Updated 2 years ago
- ATLAS tactics, techniques, and case studies data☆99Updated last month
- Practical examples of "Flawed Machine Learning Security" together with ML Security best practice across the end to end stages of the mach…☆124Updated 3 years ago
- Secure Jupyter Notebooks and Experimentation Environment☆84Updated 11 months ago
- An open-source compliance-centered evaluation framework for Generative AI models☆178Updated last month
- Make your GenAI Apps Safe & Secure Test & harden your system prompt☆610Updated last week
- Codebase of https://arxiv.org/abs/2410.14923☆54Updated last year
- Security and compliance proxy for LLM APIs☆50Updated 2 years ago
- Code for the paper "Defeating Prompt Injections by Design"☆220Updated 7 months ago
- source for llmsec.net☆16Updated last year
- This repository contains various attack against Large Language Models.☆126Updated last year
- The Privacy Adversarial Framework (PAF) is a knowledge base of privacy-focused adversarial tactics and techniques. PAF is heavily inspire…☆57Updated 2 years ago
- A guide to LLM hacking: fundamentals, prompt injection, offense, and defense☆182Updated 2 years ago
- An environment for testing AI agents against networks using Metasploit.☆45Updated 2 years ago