ZZZ-zero / JailbreakBench
View external linksLinks

🚀 JailbreakBench 是一个用于评估大语言模型(LLM)安全性的测试工具,专注于检测模型对越狱攻击(Jailbreak)的抵抗能力。通过模拟恶意提示词注入、编码攻击和多轮对话操控,量化模型的漏洞风险,并生成详细报告与可视化分析。支持中英文数据集,适用于安全研究、模型优化和合规性测试。
28Sep 1, 2025Updated 5 months ago

Alternatives and similar repositories for JailbreakBench

Users that are interested in JailbreakBench are comparing it to the libraries listed below

Sorting:

Are these results useful?