[ICML 2025] UDora: A Unified Red Teaming Framework against LLM Agents
☆32Jun 24, 2025Updated 8 months ago
Alternatives and similar repositories for UDora
Users that are interested in UDora are comparing it to the libraries listed below
Sorting:
- ☆17Jun 18, 2025Updated 8 months ago
- Pytorch implementation of NPAttack☆12Jul 7, 2020Updated 5 years ago
- this is for the ACM MM paper---Backdoor Attack on Crowd Counting☆17Jul 10, 2022Updated 3 years ago
- Code for Boosting fast adversarial training with learnable adversarial initialization (TIP2022)☆29Aug 22, 2023Updated 2 years ago
- Official implementation of the WASP web agent security benchmark☆70Aug 12, 2025Updated 6 months ago
- Code for our NAACL2025 accepted paper: Attention Tracker: Detecting Prompt Injection Attacks in LLMs☆23Sep 19, 2025Updated 5 months ago
- RAB: Provable Robustness Against Backdoor Attacks☆39Oct 3, 2023Updated 2 years ago
- Pytorch Dataloader - GTSRB (German Traffic Sign Recognition)☆20Apr 16, 2019Updated 6 years ago
- Improving fast adversarial training with prior-guided knowledge (TPAMI2024)☆43Apr 21, 2024Updated last year
- ☆20May 6, 2022Updated 3 years ago
- Code repo of our paper Towards Understanding Jailbreak Attacks in LLMs: A Representation Space Analysis (https://arxiv.org/abs/2406.10794…☆23Jul 26, 2024Updated last year
- A Unified Benchmark and Toolbox for Multimodal Jailbreak Attack–Defense Evaluation☆58Jan 23, 2026Updated last month
- This is the implementation of our paper 'Open-sourced Dataset Protection via Backdoor Watermarking', accepted by the NeurIPS Workshop on …☆23Oct 13, 2021Updated 4 years ago
- Comprehensive Assessment of Trustworthiness in Multimodal Foundation Models☆27Mar 15, 2025Updated 11 months ago
- Adversarial Attacks against Closed-Source MLLMs via Feature Optimal Alignment (NeurIPS 2025)☆49Nov 5, 2025Updated 3 months ago
- This repository contains the official code for the paper: "Prompt Injection: Parameterization of Fixed Inputs"☆32Sep 13, 2024Updated last year
- Code for Prior-Guided Adversarial Initialization for Fast Adversarial Training (ECCV2022)☆28Nov 25, 2022Updated 3 years ago
- The official PyTorch implementation of "Adversarially-Aware Robust Object Detector"☆64Sep 22, 2022Updated 3 years ago
- [CCS 2021] TSS: Transformation-specific smoothing for robustness certification☆26Oct 3, 2023Updated 2 years ago
- ☆29Mar 3, 2021Updated 4 years ago
- Vstream - Video Analytics pipeline with Hardware based accelerations (dev - stage)☆10Feb 2, 2024Updated 2 years ago
- ☆31Apr 8, 2020Updated 5 years ago
- Code repository for the paper --- [USENIX Security 2023] Towards A Proactive ML Approach for Detecting Backdoor Poison Samples☆30Jul 11, 2023Updated 2 years ago
- A curated list of papers for the transferability of adversarial examples☆76Jul 8, 2024Updated last year
- [CVPR 2022 oral] Subspace Adversarial Training☆28Apr 27, 2023Updated 2 years ago
- Code for the paper Boosting Accuracy and Robustness of Student Models via Adaptive Adversarial Distillation (CVPR 2023).☆34May 26, 2023Updated 2 years ago
- [CCS 2024] Optimization-based Prompt Injection Attack to LLM-as-a-Judge☆39Sep 17, 2025Updated 5 months ago
- Stochastic Weight Averaging Tutorials using pytorch.☆33Oct 23, 2020Updated 5 years ago
- ☆37Oct 2, 2024Updated last year
- [ICLR 2022 official code] Robust Learning Meets Generative Models: Can Proxy Distributions Improve Adversarial Robustness?☆29Mar 15, 2022Updated 3 years ago
- ☆12May 6, 2022Updated 3 years ago
- A Framework for Evaluating AI Agent Safety in Realistic Environments☆30Oct 2, 2025Updated 5 months ago
- TI-RSLK_小车走迷宫☆14Apr 27, 2019Updated 6 years ago
- The Oyster series is a set of safety models developed in-house by Alibaba-AAIG, devoted to building a responsible AI ecosystem. | Oyster …☆59Sep 11, 2025Updated 5 months ago
- ☆43Feb 9, 2026Updated 3 weeks ago
- A Multi-Session and Multi-Therapy Benchmark for High-Realism AI Psychological Counselor☆29Jan 13, 2026Updated last month
- [ICLR 2022] "Sparsity Winning Twice: Better Robust Generalization from More Efficient Training" by Tianlong Chen*, Zhenyu Zhang*, Pengjun…☆40Mar 20, 2022Updated 3 years ago
- Implementation of the paper "MAZE: Data-Free Model Stealing Attack Using Zeroth-Order Gradient Estimation".☆31Dec 12, 2021Updated 4 years ago
- Towards Memorization-Free Diffusion Models (CVPR2024) Codebase☆12Jun 2, 2024Updated last year