A framework to evaluate the generalization capability of safety alignment for LLMs
☆624Oct 9, 2025Updated 4 months ago
Alternatives and similar repositories for CipherChat
Users that are interested in CipherChat are comparing it to the libraries listed below
Sorting:
- A novel approach to improve the safety of large language models, enabling them to transition effectively from unsafe to safe state.☆71May 22, 2025Updated 9 months ago
- ☆28Mar 20, 2024Updated last year
- [ICLR 2024] The official implementation of our ICLR2024 paper "AutoDAN: Generating Stealthy Jailbreak Prompts on Aligned Large Language M…☆430Jan 22, 2025Updated last year
- ☆698Jul 2, 2025Updated 7 months ago
- ☆195Nov 26, 2023Updated 2 years ago
- Official repo for GPTFUZZER : Red Teaming Large Language Models with Auto-Generated Jailbreak Prompts☆566Sep 24, 2024Updated last year
- [ICLR 2024]Data for "Multilingual Jailbreak Challenges in Large Language Models"☆98Mar 7, 2024Updated last year
- We jailbreak GPT-3.5 Turbo’s safety guardrails by fine-tuning it on only 10 adversarially designed examples, at a cost of less than $0.20…☆339Feb 23, 2024Updated 2 years ago
- The official implementation of our NAACL 2024 paper "A Wolf in Sheep’s Clothing: Generalized Nested Jailbreak Prompts can Fool Large Lang…☆153Sep 2, 2025Updated 5 months ago
- Official implementation of our IWSLT 2023 paper "The MineTrans Systems for IWSLT 2023 Offline Speech Translation and Speech-to-Speech Tra…☆16Jul 14, 2023Updated 2 years ago
- The repo for paper: Exploiting the Index Gradients for Optimization-Based Jailbreaking on Large Language Models.☆13Dec 16, 2024Updated last year
- Universal and Transferable Attacks on Aligned Language Models☆4,521Aug 2, 2024Updated last year
- ☆121Feb 3, 2025Updated last year
- An easy-to-use Python framework to generate adversarial jailbreak prompts.☆815Mar 27, 2025Updated 11 months ago
- Improved Few-Shot Jailbreaking Can Circumvent Aligned Language Models and Their Defenses (NeurIPS 2024)☆65Jan 11, 2025Updated last year
- [ICSE'25] Aligning the Objective of LLM-based Program Repair☆23Mar 8, 2025Updated 11 months ago
- JailbreakBench: An Open Robustness Benchmark for Jailbreaking Language Models [NeurIPS 2024 Datasets and Benchmarks Track]☆531Apr 4, 2025Updated 10 months ago
- [ACL 2024] CodeAttack: Revealing Safety Generalization Challenges of Large Language Models via Code Completion☆58Oct 1, 2025Updated 4 months ago
- SpyGame: An interactive multi-agent framework to evaluate intelligence with large language models :D☆15Nov 9, 2023Updated 2 years ago
- Code and data for the paper: On the Humanity of Conversational AI: Evaluating the Psychological Portrayal of LLMs☆128Jan 24, 2026Updated last month
- Multilingual safety benchmark for Large Language Models☆53Sep 1, 2024Updated last year
- [arXiv:2311.03191] "DeepInception: Hypnotize Large Language Model to Be Jailbreaker"☆173Feb 20, 2024Updated 2 years ago
- Towards Safe LLM with our simple-yet-highly-effective Intention Analysis Prompting☆20Mar 25, 2024Updated last year
- [NeurIPS'25] Official Implementation of RISE (Reinforcing Reasoning with Self-Verification)☆31Aug 8, 2025Updated 6 months ago
- Papers and resources related to the security and privacy of LLMs 🤖☆566Jun 8, 2025Updated 8 months ago
- Recent papers on (1) Psychology of LLMs; (2) Biases in LLMs.☆50Nov 3, 2023Updated 2 years ago
- Jailbreaking Leading Safety-Aligned LLMs with Simple Adaptive Attacks [ICLR 2025]☆377Jan 23, 2025Updated last year
- Fine-tuning base models to build robust task-specific models☆34Apr 11, 2024Updated last year
- MTTM: Metamorphic Testing for Textual Content Moderation Software☆32Feb 10, 2023Updated 3 years ago
- A curated list of safety-related papers, articles, and resources focused on Large Language Models (LLMs). This repository aims to provide…☆1,772Feb 1, 2026Updated 3 weeks ago
- This is the starter kit for the Trojan Detection Challenge 2023 (LLM Edition), a NeurIPS 2023 competition.☆90May 19, 2024Updated last year
- Official Repository for ACL 2024 Paper SafeDecoding: Defending against Jailbreak Attacks via Safety-Aware Decoding☆151Jul 19, 2024Updated last year
- Code of "Improving Machine Translation with Human Feedback: An Exploration of Quality Estimation as a Reward Model"☆23Jun 28, 2024Updated last year
- [AAAI'25 (Oral)] Jailbreaking Large Vision-language Models via Typographic Visual Prompts☆191Jun 26, 2025Updated 8 months ago
- PromptInject is a framework that assembles prompts in a modular fashion to provide a quantitative analysis of the robustness of LLMs to a…☆455Feb 26, 2024Updated 2 years ago
- ☆14Feb 26, 2025Updated last year
- Official Code Implementation for the CCS 2022 Paper "On the Privacy Risks of Cell-Based NAS Architectures"☆11Nov 21, 2022Updated 3 years ago
- ☆11Jan 19, 2025Updated last year
- A curation of awesome tools, documents and projects about LLM Security.☆1,530Aug 20, 2025Updated 6 months ago