yunwei37 / prompt-hacker-collectionsLinks
prompt attack-defense, prompt Injection, reverse engineering notes and examples | 提示词对抗、破解例子与笔记
☆183Updated 3 months ago
Alternatives and similar repositories for prompt-hacker-collections
Users that are interested in prompt-hacker-collections are comparing it to the libraries listed below
Sorting:
- Official repo for GPTFUZZER : Red Teaming Large Language Models with Auto-Generated Jailbreak Prompts☆494Updated 8 months ago
- This repo includes ChatGPT jailbreaking prompts and some tests on its perference☆72Updated last year
- Hack OpenAI LLMs' System Prompts By Reverse Prompt Engineering☆66Updated 2 years ago
- CS-Eval is a comprehensive evaluation suite for fundamental cybersecurity models or large language models' cybersecurity ability.☆43Updated 6 months ago
- This repository provides a benchmark for prompt Injection attacks and defenses☆216Updated this week
- GPT-Analyst: A GPT for GPT analysis and reverse engineering☆199Updated last year
- ☆71Updated 6 months ago
- The official implementation of our NAACL 2024 paper "A Wolf in Sheep’s Clothing: Generalized Nested Jailbreak Prompts can Fool Large Lang…☆111Updated 4 months ago
- PromptInject is a framework that assembles prompts in a modular fashion to provide a quantitative analysis of the robustness of LLMs to a…☆374Updated last year
- A curated list of awesome security tools, experimental case or other interesting things with LLM or GPT.☆598Updated last week
- ☆134Updated 9 months ago
- 这个资源库包含了为 Prompt 工程手工整理的资源中文清单,重点是GPT、ChatGPT、PaLM 等(自动持续更新)☆127Updated 2 years ago
- Persuasive Jailbreaker: we can persuade LLMs to jailbreak them!☆302Updated 7 months ago
- Papers and resources related to the security and privacy of LLMs 🤖☆504Updated 6 months ago
- Protect your GPTs through secure prompts to prevent malicious data leaks. 通过安全 Prompt 保护你的 GPTs ,避免内容被恶意泄露☆185Updated 5 months ago
- ShieldLM: Empowering LLMs as Aligned, Customizable and Explainable Safety Detectors [EMNLP 2024 Findings]☆191Updated 8 months ago
- [NDSS'25 Best Technical Poster] A collection of automated evaluators for assessing jailbreak attempts.☆158Updated 2 months ago
- TAP: An automated jailbreaking method for black-box LLMs☆171Updated 5 months ago
- ☆24Updated 7 months ago
- LLM security and privacy☆49Updated 7 months ago
- An easy-to-use Python framework to generate adversarial jailbreak prompts.☆649Updated 2 months ago
- 🤖🛡️🔍🔒🔑 Tiny package designed to support red teams and penetration testers in exploiting large language model AI solutions.☆23Updated last year
- [ICLR 2024] The official implementation of our ICLR2024 paper "AutoDAN: Generating Stealthy Jailbreak Prompts on Aligned Large Language M…☆334Updated 4 months ago
- ⚡ Vigil ⚡ Detect prompt injections, jailbreaks, and other potentially risky Large Language Model (LLM) inputs☆389Updated last year
- 2024 Alibaba Global Mathematics Competition AI Track Global 2nd Place Project (Agent Universe)☆68Updated 11 months ago
- SC-Safety: 中文大模型多轮对抗安全基准☆135Updated last year
- Bag of Tricks: Benchmarking of Jailbreak Attacks on LLMs. Empirical tricks for LLM Jailbreaking. (NeurIPS 2024)☆135Updated 6 months ago
- [arXiv:2311.03191] "DeepInception: Hypnotize Large Language Model to Be Jailbreaker"☆145Updated last year
- 微软 AutoGen 框架 Demo☆83Updated last year
- Code for the website www.jailbreakchat.com☆95Updated last year