alohachen / Hide-and-Seek
Hide and Seek (HaS): A Framework for Prompt Privacy Protection
☆30Updated last year
Alternatives and similar repositories for Hide-and-Seek:
Users that are interested in Hide-and-Seek are comparing it to the libraries listed below
- ShieldLM: Empowering LLMs as Aligned, Customizable and Explainable Safety Detectors [EMNLP 2024 Findings]☆170Updated 3 months ago
- EmojiCrypt: Prompt Encryption for Secure Communication with Large Language Models☆13Updated 10 months ago
- A survey of privacy problems in Large Language Models (LLMs). Contains summary of the corresponding paper along with relevant code☆65Updated 7 months ago
- The code for paper "The Good and The Bad: Exploring Privacy Issues in Retrieval-Augmented Generation (RAG)", exploring the privacy risk o…☆39Updated last month
- ☆55Updated 2 weeks ago
- [NAACL2024] Attacks, Defenses and Evaluations for LLM Conversation Safety: A Survey☆86Updated 5 months ago
- Chain of Attack: a Semantic-Driven Contextual Multi-Turn attacker for LLM☆28Updated last week
- A toolkit to assess data privacy in LLMs (under development)☆47Updated 2 weeks ago
- ☆66Updated 2 months ago
- [USENIX Security 2025] PoisonedRAG: Knowledge Corruption Attacks to Retrieval-Augmented Generation of Large Language Models☆108Updated 3 months ago
- S-Eval: Automatic and Adaptive Test Generation for Benchmarking Safety Evaluation of Large Language Models☆47Updated 2 months ago
- ☆37Updated 7 months ago
- ☆106Updated 4 months ago
- Code for Findings-EMNLP 2023 paper: Multi-step Jailbreaking Privacy Attacks on ChatGPT☆29Updated last year
- ☆11Updated 10 months ago
- The repository contains the code for analysing the leakage of personally identifiable (PII) information from the output of next word pred…☆88Updated 5 months ago
- ☆69Updated 2 years ago
- Official github repo for SafetyBench, a comprehensive benchmark to evaluate LLMs' safety. [ACL 2024]☆182Updated 6 months ago
- Privacy-Preserving Prompt Tuning for Large Language Model☆18Updated 9 months ago
- Shepherd: A foundational framework enabling federated instruction tuning for large language models☆216Updated last year
- ☆15Updated this week
- JailBench:大型语言模型越狱攻击风险评测中文数据集☆31Updated 6 months ago
- The official implementation of our NAACL 2024 paper "A Wolf in Sheep’s Clothing: Generalized Nested Jailbreak Prompts can Fool Large Lang…☆88Updated last month
- LLM Unlearning☆141Updated last year
- Official repository for ICML 2024 paper "On Prompt-Driven Safeguarding for Large Language Models"☆83Updated 4 months ago
- Federated Learning for LLMs.☆180Updated last month
- ☆48Updated 3 weeks ago
- Code&Data for the paper "Watch Out for Your Agents! Investigating Backdoor Threats to LLM-Based Agents" [NeurIPS 2024]☆60Updated 3 months ago
- BackdoorLLM: A Comprehensive Benchmark for Backdoor Attacks on Large Language Models☆94Updated this week
- R-Judge: Benchmarking Safety Risk Awareness for LLM Agents (EMNLP Findings 2024)☆65Updated 3 months ago