A toolbox for benchmarking Multimodal LLM Agents trustworthiness across truthfulness, controllability, safety and privacy dimensions through 34 interactive tasks
☆64Jan 9, 2026Updated last month
Alternatives and similar repositories for MLA-Trust
Users that are interested in MLA-Trust are comparing it to the libraries listed below
Sorting:
- Official implementation of the paper: ANF: Crafting transferable adversarial point clouds via adversarial noise factorization☆30Oct 6, 2025Updated 5 months ago
- On the Robustness of GUI Grounding Models Against Image Attacks☆12Apr 8, 2025Updated 10 months ago
- [ICML 2025] X-Transfer Attacks: Towards Super Transferable Adversarial Attacks on CLIP☆39Feb 3, 2026Updated last month
- ☆14Feb 26, 2025Updated last year
- [NDSS'25] The official implementation of safety misalignment.☆17Jan 8, 2025Updated last year
- [CVPR2024 Highlight] Strong Transferable Adversarial Attacks via Ensembled Asymptotically Normal Distribution Learning☆19Jun 14, 2024Updated last year
- SG-Bench: Evaluating LLM Safety Generalization Across Diverse Tasks and Prompt Types☆25Nov 29, 2024Updated last year
- ☆21Mar 17, 2025Updated 11 months ago
- Official implementation of the paper: Enhancing the Transferability of Adversarial Point Clouds by Initializing Transferable Adversarial …☆29Apr 1, 2025Updated 11 months ago
- [ICCV-2025] Universal Adversarial Attack, Multimodal Adversarial Attacks, VLP models, Contrastive Learning, Cross-modal Perturbation Gene…☆35Jul 10, 2025Updated 7 months ago