llm-platform-security / chatgpt-plugin-evalLinks
LLM Platform Security: Applying a Systematic Evaluation Framework to OpenAI's ChatGPT Plugins
☆25Updated 10 months ago
Alternatives and similar repositories for chatgpt-plugin-eval
Users that are interested in chatgpt-plugin-eval are comparing it to the libraries listed below
Sorting:
- ☆24Updated 3 years ago
- ☆114Updated 10 months ago
- ☆66Updated 4 years ago
- 🔥🔥🔥 Detecting hidden backdoors in Large Language Models with only black-box access☆29Updated this week
- Machine Learning & Security Seminar @Purdue University☆25Updated 2 years ago
- Code used to run the platform for the LLM CTF colocated with SaTML 2024☆26Updated last year
- ☆16Updated 9 months ago
- An Execution Isolation Architecture for LLM-Based Agentic Systems☆80Updated 4 months ago
- Unofficial implementation of "Backdooring Instruction-Tuned Large Language Models with Virtual Prompt Injection"☆18Updated 11 months ago
- This repository provides a benchmark for prompt Injection attacks and defenses☆216Updated this week
- ☆57Updated 5 months ago
- Code&Data for the paper "Watch Out for Your Agents! Investigating Backdoor Threats to LLM-Based Agents" [NeurIPS 2024]☆79Updated 8 months ago
- Learning Security Classifiers with Verified Global Robustness Properties (CCS'21) https://arxiv.org/pdf/2105.11363.pdf☆28Updated 3 years ago
- Code to conduct an embedding attack on LLMs☆25Updated 4 months ago
- Agent Security Bench (ASB)☆81Updated last month
- LLM security and privacy☆49Updated 7 months ago
- ☆19Updated last year
- ☆63Updated 11 months ago
- The official implementation of our pre-print paper "Automatic and Universal Prompt Injection Attacks against Large Language Models".☆49Updated 7 months ago
- ☆9Updated 4 years ago
- ☆43Updated 8 months ago
- PAL: Proxy-Guided Black-Box Attack on Large Language Models☆51Updated 9 months ago
- Repository for "SecurityEval Dataset: Mining Vulnerability Examples to Evaluate Machine Learning-Based Code Generation Techniques" publis…☆71Updated last year
- [USENIX Security '24] An LLM-Assisted Easy-to-Trigger Backdoor Attack on Code Completion Models: Injecting Disguised Vulnerabilities agai…☆45Updated 2 months ago
- Fine-tuning base models to build robust task-specific models☆30Updated last year
- The repository contains the code for analysing the leakage of personally identifiable (PII) information from the output of next word pred…☆96Updated 9 months ago
- Official implementation of paper: DrAttack: Prompt Decomposition and Reconstruction Makes Powerful LLM Jailbreakers☆52Updated 9 months ago
- Bad Characters: Imperceptible NLP Attacks☆34Updated last year
- A curated list of trustworthy Generative AI papers. Daily updating...☆73Updated 9 months ago
- Explainable malware and vulnerability detection with XAI in paper "FINER: Enhancing State-of-the-art Classifiers with Feature Attribution…☆11Updated 9 months ago