gitkolento / SecProbe
SecProbe:任务驱动式大模型安全能力评测系统
☆13Updated 4 months ago
Alternatives and similar repositories for SecProbe:
Users that are interested in SecProbe are comparing it to the libraries listed below
- ☆13Updated 2 months ago
- JailBench:大型语言模型越狱攻击风险评测中文数据集 [PAKDD 2025]☆81Updated last month
- 针对大语言模型的对抗性攻击总结☆25Updated last year
- ☆22Updated 6 months ago
- Code for paper "The Philosopher’s Stone: Trojaning Plugins of Large Language Models"☆16Updated 7 months ago
- ☆79Updated last year
- 复旦白泽大模型安全基准测试集(2024年夏季版)☆36Updated 8 months ago
- Safety at Scale: A Comprehensive Survey of Large Model Safety☆149Updated 2 months ago
- "他山之石、可以攻玉":复旦白泽智能发布面向国内开源和国外商用大模型的Demo数据集JADE-DB☆399Updated last month
- ☆9Updated 6 months ago
- ☆81Updated 2 months ago
- ☆51Updated 4 months ago
- Official github repo for SafetyBench, a comprehensive benchmark to evaluate LLMs' safety. [ACL 2024]☆213Updated 10 months ago
- [USENIX Security 2025] PoisonedRAG: Knowledge Corruption Attacks to Retrieval-Augmented Generation of Large Language Models☆141Updated 2 months ago
- A curated list of safety-related papers, articles, and resources focused on Large Language Models (LLMs). This repository aims to provide…☆1,346Updated this week
- ☆17Updated 2 months ago
- BackdoorLLM: A Comprehensive Benchmark for Backdoor Attacks on Large Language Models☆134Updated 2 months ago
- This is the code repository for "Uncovering Safety Risks of Large Language Models through Concept Activation Vector"☆36Updated 5 months ago
- [USENIX Security'24] Official repository of "Making Them Ask and Answer: Jailbreaking Large Language Models in Few Queries via Disguise a…☆86Updated 6 months ago
- Chain of Attack: a Semantic-Driven Contextual Multi-Turn attacker for LLM☆29Updated 3 months ago
- Agent Security Bench (ASB)☆76Updated 3 weeks ago
- ☆46Updated 10 months ago
- ☆128Updated 7 months ago
- [ICLR 2024] The official implementation of our ICLR2024 paper "AutoDAN: Generating Stealthy Jailbreak Prompts on Aligned Large Language M…☆319Updated 3 months ago
- Accepted by ECCV 2024☆125Updated 6 months ago
- [NeurIPS 2024] Official implementation for "AgentPoison: Red-teaming LLM Agents via Memory or Knowledge Base Backdoor Poisoning"☆116Updated 2 weeks ago
- Bag of Tricks: Benchmarking of Jailbreak Attacks on LLMs. Empirical tricks for LLM Jailbreaking. (NeurIPS 2024)☆135Updated 4 months ago
- JAILJUDGE: A comprehensive evaluation benchmark which includes a wide range of risk scenarios with complex malicious prompts (e.g., synth…☆44Updated 4 months ago
- Awesome Large Reasoning Model(LRM) Safety.This repository is used to collect security-related research on large reasoning models such as …☆63Updated this week
- An easy-to-use Python framework to defend against jailbreak prompts.☆20Updated last month