THU-BPM / MarkLLMLinks
MarkLLM: An Open-Source Toolkit for LLM Watermarking.οΌEMNLP 2024 System Demonstration)
β627Updated last week
Alternatives and similar repositories for MarkLLM
Users that are interested in MarkLLM are comparing it to the libraries listed below
Sorting:
- UP-TO-DATE LLM Watermark paper. π₯π₯π₯β357Updated 10 months ago
- AISafetyLab: A comprehensive framework covering safety attack, defense, evaluation and paper list.β203Updated last month
- [ICLR 2025 Spotlight] The official implementation of our ICLR2025 paper "AutoDAN-Turbo: A Lifelong Agent for Strategy Self-Exploration toβ¦β310Updated this week
- β39Updated last year
- β149Updated last year
- π up-to-date & curated list of awesome Attacks on Large-Vision-Language-Models papers, methods & resources.β394Updated 2 weeks ago
- [USENIX Security 2025] PoisonedRAG: Knowledge Corruption Attacks to Retrieval-Augmented Generation of Large Language Modelsβ202Updated 7 months ago
- Improved techniques for optimization-based jailbreaking on large language models (ICLR2025)β130Updated 6 months ago
- Papers and resources related to the security and privacy of LLMs π€β536Updated 4 months ago
- An easy-to-use Python framework to generate adversarial jailbreak prompts.β724Updated 6 months ago
- multi-bit language model watermarking (NAACL 24)β15Updated last year
- β135Updated 7 months ago
- Repository for Towards Codable Watermarking for Large Language Modelsβ38Updated 2 years ago
- A survey on harmful fine-tuning attack for large language modelβ212Updated this week
- This is the code repository for "Uncovering Safety Risks of Large Language Models through Concept Activation Vector"β44Updated 10 months ago
- [CCS'24] SafeGen: Mitigating Unsafe Content Generation in Text-to-Image Modelsβ136Updated 3 months ago
- β629Updated 3 weeks ago
- Code and data for paper "A Semantic Invariant Robust Watermark for Large Language Models" accepted by ICLR 2024.β34Updated 10 months ago
- Safety at Scale: A Comprehensive Survey of Large Model Safetyβ194Updated 7 months ago
- Source code of paper "An Unforgeable Publicly Verifiable Watermark for Large Language Models" accepted by ICLR 2024β35Updated last year
- β44Updated 7 months ago
- A curated list of resources dedicated to the safety of Large Vision-Language Models. This repository aligns with our survey titled A Survβ¦β149Updated this week
- β¨ A synthetic dataset generation framework that produces diverse coding questions and verifiable solutions - all in one framworkβ276Updated last month
- YiJian-Comunity: a full-process automated large model safety evaluation tool designed for academic researchβ114Updated last year
- Accepted by IJCAI-24 Survey Trackβ216Updated last year
- Awesome-Jailbreak-on-LLMs is a collection of state-of-the-art, novel, exciting jailbreak methods on LLMs. It contains papers, codes, dataβ¦β939Updated last month
- β16Updated 5 months ago
- [NAACL2024] Attacks, Defenses and Evaluations for LLM Conversation Safety: A Surveyβ106Updated last year
- This is the repo for the paper "OS Agents: A Survey on MLLM-based Agents for Computer, Phone and Browser Use" (ACL 2025 Oral).β354Updated last month
- This repository contains the source code, datasets, and scripts for the paper "GenderCARE: A Comprehensive Framework for Assessing and Reβ¦β24Updated last year