ZenGuard-AI / fast-llm-security-guardrails
The fastest && easiest LLM security guardrails for AI Agents and applications.
☆101Updated 2 weeks ago
Related projects ⓘ
Alternatives and complementary repositories for fast-llm-security-guardrails
- Red-Teaming Language Models with DSPy☆142Updated 7 months ago
- ☆34Updated 3 months ago
- A trace analysis tool for AI agents.☆124Updated last month
- Framework for LLM evaluation, guardrails and security☆96Updated 2 months ago
- Jailbreaking Leading Safety-Aligned LLMs with Simple Adaptive Attacks [arXiv, Apr 2024]☆220Updated 2 months ago
- LLM security and privacy☆41Updated last month
- Payloads for Attacking Large Language Models☆64Updated 4 months ago
- ☆61Updated last month
- The Rule-based Retrieval package is a Python package that enables you to create and manage Retrieval Augmented Generation (RAG) applicati…☆222Updated last month
- SecGPT: An execution isolation architecture for LLM-based systems☆49Updated 3 weeks ago
- Risks and targets for assessing LLMs & LLM vulnerabilities☆25Updated 5 months ago
- A repository of Language Model Vulnerabilities and Exposures (LVEs).☆107Updated 8 months ago
- Sphynx Hallucination Induction☆48Updated 3 months ago
- Fiddler Auditor is a tool to evaluate language models.☆171Updated 8 months ago
- [ICLR'24 Spotlight] A language model (LM)-based emulation framework for identifying the risks of LM agents with tool use☆115Updated 8 months ago
- This repository provides implementation to formalize and benchmark Prompt Injection attacks and defenses☆146Updated 2 months ago
- A benchmark for prompt injection detection systems.☆87Updated 2 months ago
- Do-Not-Answer: A Dataset for Evaluating Safeguards in LLMs☆184Updated 5 months ago
- Tutorial for building LLM router☆163Updated 4 months ago
- A library for red-teaming LLM applications with LLMs.☆22Updated last month
- Self-hardening firewall for large language models☆258Updated 8 months ago
- ☆63Updated this week
- LLM | Security | Operations in one github repo with good links and pictures.☆19Updated last month
- A subset of jailbreaks automatically discovered by the Haize Labs haizing suite.☆86Updated 5 months ago
- Official repository for the paper "ALERT: A Comprehensive Benchmark for Assessing Large Language Models’ Safety through Red Teaming"☆33Updated 2 months ago
- Curation of prompts that are known to be adversarial to large language models☆174Updated last year
- Research on "Many-Shot Jailbreaking" in Large Language Models (LLMs). It unveils a novel technique capable of bypassing the safety mechan…☆18Updated 7 months ago
- TAP: An automated jailbreaking method for black-box LLMs☆119Updated 8 months ago
- Papers about red teaming LLMs and Multimodal models.☆78Updated this week
- Official repo for Customized but Compromised: Assessing Prompt Injection Risks in User-Designed GPTs☆21Updated last year