ThuCCSLab / Awesome-LM-SSP
A reading list for large models safety, security, and privacy (including Awesome LLM Security, Safety, etc.).
☆750Updated this week
Related projects: ⓘ
- A curated list of safety-related papers, articles, and resources focused on Large Language Models (LLMs). This repository aims to provide…☆844Updated this week
- Papers and resources related to the security and privacy of LLMs 🤖☆393Updated last week
- A curation of awesome tools, documents and projects about LLM Security.☆873Updated 3 weeks ago
- Accepted by IJCAI-24 Survey Track☆117Updated 3 weeks ago
- [ICML 2024] TrustLLM: Trustworthiness in Large Language Models☆432Updated 2 weeks ago
- 😎 up-to-date & curated list of awesome Attacks on Large-Vision-Language-Models papers, methods & resources.☆73Updated this week
- UP-TO-DATE LLM Watermark paper. 🔥🔥🔥☆253Updated 3 months ago
- A curated list of Meachine learning Security & Privacy papers published in security top-4 conferences (IEEE S&P, ACM CCS, USENIX Security…☆200Updated last month
- The official implementation of our ICLR2024 paper "AutoDAN: Generating Stealthy Jailbreak Prompts on Aligned Large Language Models".☆203Updated last month
- Must-read Papers on Knowledge Editing for Large Language Models.☆829Updated 2 weeks ago
- MarkLLM: An Open-Source Toolkit for LLM Watermarking.☆246Updated last month
- An Awesome Collection for LLM Survey☆289Updated last week
- Repository for the Paper (AAAI 2024, Oral) --- Visual Adversarial Examples Jailbreak Large Language Models☆156Updated 4 months ago
- LLM hallucination paper list☆268Updated 6 months ago
- An easy-to-use Python framework to generate adversarial jailbreak prompts.☆403Updated 2 weeks ago
- A curated list of papers & resources linked to data poisoning, backdoor attacks and defenses against them (no longer maintained)☆170Updated 2 months ago
- Reading list of hallucination in LLMs. Check out our new survey paper: "Siren’s Song in the AI Ocean: A Survey on Hallucination in Large …☆901Updated 2 weeks ago
- A curated list of papers & resources on backdoor attacks and defenses in deep learning.☆165Updated 6 months ago
- The lastest paper about detection of LLM-generated text and code☆195Updated last week
- ☆381Updated last month
- We jailbreak GPT-3.5 Turbo’s safety guardrails by fine-tuning it on only 10 adversarially designed examples, at a cost of less than $0.20…☆219Updated 6 months ago
- [NAACL2024] Attacks, Defenses and Evaluations for LLM Conversation Safety: A Survey☆65Updated last month
- ☆269Updated 2 months ago
- Awesome-LLM-Robustness: a curated list of Uncertainty, Reliability and Robustness in Large Language Models☆627Updated 3 months ago
- ☆207Updated 3 months ago
- Jailbreaking Large Vision-language Models via Typographic Visual Prompts☆76Updated 4 months ago
- An Open Robustness Benchmark for Jailbreaking Language Models [arXiv 2024]☆169Updated last month
- Continuously updated list of related resources for generative LLMs like GPT and their analysis and detection.☆187Updated 2 weeks ago
- A resource repository for machine unlearning in large language models☆131Updated this week
- ☆357Updated last month