theshi-1128 / llm-defenseLinks
An easy-to-use Python framework to defend against jailbreak prompts.
☆21Updated 7 months ago
Alternatives and similar repositories for llm-defense
Users that are interested in llm-defense are comparing it to the libraries listed below
Sorting:
- Safety at Scale: A Comprehensive Survey of Large Model Safety☆200Updated 8 months ago
- ☆66Updated 5 months ago
- [USENIX Security 2025] PoisonedRAG: Knowledge Corruption Attacks to Retrieval-Augmented Generation of Large Language Models☆210Updated 8 months ago
- ☆82Updated last month
- [AAAI'25 (Oral)] Jailbreaking Large Vision-language Models via Typographic Visual Prompts☆177Updated 4 months ago
- ☆36Updated last year
- An open-source toolkit for textual backdoor attack and defense (NeurIPS 2022 D&B, Spotlight)☆196Updated 2 years ago
- [NeurIPS 2025] BackdoorLLM: A Comprehensive Benchmark for Backdoor Attacks and Defenses on Large Language Models