alexalbertt / jailbreakchatLinks
Code for the website www.jailbreakchat.com
☆97Updated last year
Alternatives and similar repositories for jailbreakchat
Users that are interested in jailbreakchat are comparing it to the libraries listed below
Sorting:
- PromptInject is a framework that assembles prompts in a modular fashion to provide a quantitative analysis of the robustness of LLMs to a…☆393Updated last year
- Persuasive Jailbreaker: we can persuade LLMs to jailbreak them!☆308Updated 9 months ago
- Curation of prompts that are known to be adversarial to large language models☆180Updated 2 years ago
- ☆573Updated 2 weeks ago
- A benchmark for prompt injection detection systems.☆122Updated 2 months ago
- Jailbreaking Leading Safety-Aligned LLMs with Simple Adaptive Attacks [ICLR 2025]☆320Updated 5 months ago
- ⚡ Vigil ⚡ Detect prompt injections, jailbreaks, and other potentially risky Large Language Model (LLM) inputs☆396Updated last year
- Red-Teaming Language Models with DSPy☆202Updated 5 months ago
- A subset of jailbreaks automatically discovered by the Haize Labs haizing suite.☆92Updated 3 months ago
- This repository provides a benchmark for prompt Injection attacks and defenses☆245Updated last month
- GPT-Analyst: A GPT for GPT analysis and reverse engineering☆200Updated last year
- ComPromptMized: Unleashing Zero-click Worms that Target GenAI-Powered Applications☆203Updated last year
- Learn about a type of vulnerability that specifically targets machine learning models☆309Updated last year
- Every practical and proposed defense against prompt injection.☆495Updated 4 months ago
- Official repo for GPTFUZZER : Red Teaming Large Language Models with Auto-Generated Jailbreak Prompts☆504Updated 9 months ago
- The fastest Trust Layer for AI Agents☆138Updated last month
- Codebase of https://arxiv.org/abs/2410.14923☆48Updated 8 months ago
- LLM security and privacy☆48Updated 9 months ago
- ☆121Updated last month
- Experimental tools to backdoor large language models by re-writing their system prompts at a raw parameter level. This allows you to pote…☆176Updated 3 months ago
- A prompt injection game to collect data for robust ML research☆62Updated 5 months ago
- A Dynamic Environment to Evaluate Attacks and Defenses for LLM Agents.☆209Updated this week
- This repository contains various attack against Large Language Models.☆111Updated last year
- ☆69Updated last year
- This project investigates the security of large language models by performing binary classification of a set of input prompts to discover…☆42Updated last year
- a prompt injection scanner for custom LLM applications☆835Updated 4 months ago
- Whispers in the Machine: Confidentiality in Agentic Systems☆39Updated last month
- Moonshot - A simple and modular tool to evaluate and red-team any LLM application.☆256Updated 2 weeks ago
- A benchmark for evaluating the robustness of LLMs and defenses to indirect prompt injection attacks.☆72Updated last year
- prompt attack-defense, prompt Injection, reverse engineering notes and examples | 提示词对抗、破解例子与笔记☆196Updated 4 months ago