bailichangan / Ghidorah
☆13Updated 2 months ago
Alternatives and similar repositories for Ghidorah:
Users that are interested in Ghidorah are comparing it to the libraries listed below
- SecProbe:任务驱动式大模型安全能力评测系统☆11Updated 2 months ago
- [USENIX Security 2025] PoisonedRAG: Knowledge Corruption Attacks to Retrieval-Augmented Generation of Large Language Models☆117Updated 2 weeks ago
- 😎 up-to-date & curated list of awesome Attacks on Large-Vision-Language-Models papers, methods & resources.☆213Updated this week
- ☆38Updated last month
- ☆11Updated 11 months ago
- "他山之石、可以攻玉":复旦白泽智能发布面向国内开源和国外商用大模型的Demo数据集JADE-DB☆364Updated 2 months ago
- ☆112Updated 5 months ago
- ☆10Updated 7 months ago
- [ICLR 2024] The official implementation of our ICLR2024 paper "AutoDAN: Generating Stealthy Jailbreak Prompts on Aligned Large Language M…☆292Updated last month
- 针对大语言模型的对抗性攻击总结☆16Updated last year
- ☆41Updated 8 months ago
- ☆15Updated last month
- ☆18Updated 8 months ago
- [arXiv:2311.03191] "DeepInception: Hypnotize Large Language Model to Be Jailbreaker"☆134Updated last year
- This is the code repository for "Uncovering Safety Risks of Large Language Models through Concept Activation Vector"☆25Updated 3 months ago
- A collection of resources that investigate social agents.☆112Updated last month
- ☆74Updated 2 weeks ago
- BackdoorLLM: A Comprehensive Benchmark for Backdoor Attacks on Large Language Models☆103Updated this week
- ☆14Updated 5 months ago
- Accepted by IJCAI-24 Survey Track☆190Updated 5 months ago
- A curated list of safety-related papers, articles, and resources focused on Large Language Models (LLMs). This repository aims to provide…☆1,182Updated 2 weeks ago
- ☆13Updated last year
- This Github repository summarizes a list of research papers on AI security from the four top academic conferences.☆107Updated last year
- Bag of Tricks: Benchmarking of Jailbreak Attacks on LLMs. Empirical tricks for LLM Jailbreaking. (NeurIPS 2024)☆118Updated 2 months ago
- ☆77Updated 10 months ago
- HiSim: A Hybrid Social Media Simulation Framework☆25Updated 8 months ago
- Chain of Attack: a Semantic-Driven Contextual Multi-Turn attacker for LLM☆28Updated last month