THU-BPM / MarkLLMLinks
MarkLLM: An Open-Source Toolkit for LLM Watermarking.(EMNLP 2024 Demo)
☆418Updated last week
Alternatives and similar repositories for MarkLLM
Users that are interested in MarkLLM are comparing it to the libraries listed below
Sorting:
- UP-TO-DATE LLM Watermark paper. 🔥🔥🔥☆342Updated 6 months ago
- ☆38Updated 10 months ago
- ☆139Updated 9 months ago
- A survey on harmful fine-tuning attack for large language model☆184Updated this week
- Accepted by IJCAI-24 Survey Track☆207Updated 10 months ago
- Repository for Towards Codable Watermarking for Large Language Models☆37Updated last year
- This is the code repository for "Uncovering Safety Risks of Large Language Models through Concept Activation Vector"☆40Updated 7 months ago
- BackdoorLLM: A Comprehensive Benchmark for Backdoor Attacks and Defenses on Large Language Models☆167Updated this week
- Code for watermarking language models☆79Updated 9 months ago
- ☆594Updated last year
- multi-bit language model watermarking (NAACL 24)☆13Updated 9 months ago
- The lastest paper about detection of LLM-generated text and code☆271Updated last week
- [ACL2024-Main] Data and Code for WaterBench: Towards Holistic Evaluation of LLM Watermarks☆26Updated last year
- ShieldLM: Empowering LLMs as Aligned, Customizable and Explainable Safety Detectors [EMNLP 2024 Findings]☆198Updated 8 months ago
- [ICML 2024] TrustLLM: Trustworthiness in Large Language Models☆574Updated 3 months ago
- 😎 up-to-date & curated list of awesome Attacks on Large-Vision-Language-Models papers, methods & resources.☆330Updated this week
- Accepted by ECCV 2024☆139Updated 8 months ago
- Repo for SemStamp (NAACL2024) and k-SemStamp (ACL2024)☆20Updated 6 months ago
- Official Repository for ACL 2024 Paper SafeDecoding: Defending against Jailbreak Attacks via Safety-Aware Decoding☆134Updated 11 months ago
- [USENIX Security'24] REMARK-LLM: A robust and efficient watermarking framework for generative large language models☆25Updated 8 months ago
- [NAACL2024] Attacks, Defenses and Evaluations for LLM Conversation Safety: A Survey☆103Updated 10 months ago
- Safety at Scale: A Comprehensive Survey of Large Model Safety☆173Updated 4 months ago
- ☆29Updated 2 months ago
- ☆88Updated last month
- Continuously updated list of related resources for generative LLMs like GPT and their analysis and detection.☆222Updated last month
- Papers and resources related to the security and privacy of LLMs 🤖☆511Updated 2 weeks ago
- LLM hallucination paper list☆318Updated last year
- Official github repo for SafetyBench, a comprehensive benchmark to evaluate LLMs' safety. [ACL 2024]☆225Updated last year
- We jailbreak GPT-3.5 Turbo’s safety guardrails by fine-tuning it on only 10 adversarially designed examples, at a cost of less than $0.20…☆303Updated last year
- ☆88Updated 4 months ago