Fish-and-Sheep / Text-FluoroscopyLinks
☆12Updated 6 months ago
Alternatives and similar repositories for Text-Fluoroscopy
Users that are interested in Text-Fluoroscopy are comparing it to the libraries listed below
Sorting:
- Repository for Towards Codable Watermarking for Large Language Models☆38Updated last year
- ☆82Updated last year
- ☆35Updated 11 months ago
- Agent Security Bench (ASB)☆111Updated 2 months ago
- Safety at Scale: A Comprehensive Survey of Large Model Safety☆187Updated 6 months ago
- [USENIX Security 2025] PoisonedRAG: Knowledge Corruption Attacks to Retrieval-Augmented Generation of Large Language Models☆185Updated 6 months ago
- Code for ACM MM2024 paper: White-box Multimodal Jailbreaks Against Large Vision-Language Models☆30Updated 8 months ago
- multi-bit language model watermarking (NAACL 24)☆15Updated 11 months ago
- Robust natural language watermarking using invariant features☆26Updated last year
- ☆31Updated 5 months ago
- [USENIX Security '24] An LLM-Assisted Easy-to-Trigger Backdoor Attack on Code Completion Models: Injecting Disguised Vulnerabilities agai…☆51Updated 5 months ago
- Watermarking Text Generated by Black-Box Language Models☆39Updated last year
- enchmarking Large Language Models' Resistance to Malicious Code☆12Updated 9 months ago
- Chain of Attack: a Semantic-Driven Contextual Multi-Turn attacker for LLM☆35Updated 7 months ago
- The lastest paper about detection of LLM-generated text and code☆276Updated 2 months ago
- ☆60Updated 3 months ago
- official implementation of [USENIX Sec'25] StruQ: Defending Against Prompt Injection with Structured Queries☆46Updated last month
- An LLM can Fool Itself: A Prompt-Based Adversarial Attack (ICLR 2024)☆98Updated 7 months ago
- [AAAI'25 (Oral)] Jailbreaking Large Vision-language Models via Typographic Visual Prompts☆167Updated 2 months ago
- ☆147Updated last year
- ☆60Updated 5 months ago
- Unofficial implementation of "Backdooring Instruction-Tuned Large Language Models with Virtual Prompt Injection"☆22Updated last year
- [USENIX Security'24] REMARK-LLM: A robust and efficient watermarking framework for generative large language models☆25Updated 10 months ago
- ☆23Updated 2 years ago
- [NAACL2024] Attacks, Defenses and Evaluations for LLM Conversation Safety: A Survey☆106Updated last year
- ☆106Updated 4 months ago
- ☆101Updated 7 months ago
- S-Eval: Towards Automated and Comprehensive Safety Evaluation for Large Language Models☆95Updated 2 months ago
- Accepted by IJCAI-24 Survey Track☆212Updated last year
- Fingerprint large language models☆41Updated last year