alipay / YiJian-CommunityLinks
YiJian-Comunity: a full-process automated large model safety evaluation tool designed for academic research
☆110Updated 7 months ago
Alternatives and similar repositories for YiJian-Community
Users that are interested in YiJian-Community are comparing it to the libraries listed below
Sorting:
- [CCS'24] SafeGen: Mitigating Unsafe Content Generation in Text-to-Image Models☆130Updated 2 months ago
- Improved techniques for optimization-based jailbreaking on large language models (ICLR2025)☆102Updated 2 months ago
- [CVPR2024] MMA-Diffusion: MultiModal Attack on Diffusion Models☆161Updated last year
- [NDSS'24] Inaudible Adversarial Perturbation: Manipulating the Recognition of User Speech in Real Time☆55Updated 8 months ago
- Code for Semantic-Aligned Adversarial Evolution Triangle for High-Transferability Vision-Language Attack☆34Updated 7 months ago
- Improving fast adversarial training with prior-guided knowledge (TPAMI2024)☆41Updated last year
- A curated list of resources dedicated to the safety of Large Vision-Language Models. This repository aligns with our survey titled A Surv…☆98Updated last month
- [MM24 Oral] Identity-Driven Multimedia Forgery Detection via Reference Assistance☆103Updated last month
- [ACL 2024] User-friendly evaluation framework: Eval Suite & Benchmarks: UHGEval, HaluEval, HalluQA, etc.☆167Updated 6 months ago
- Controllable Text Generation for Large Language Models: A Survey☆175Updated 9 months ago
- ☆120Updated this week
- AISafetyLab: A comprehensive framework covering safety attack, defense, evaluation and paper list.☆179Updated 3 weeks ago
- A comprehensive collection of resources focused on addressing and understanding hallucination phenomena in MLLMs.☆34Updated last year
- A collection of multimodal reasoning papers, codes, datasets, benchmarks and resources.☆223Updated this week
- Machine-generated text detection in the wild (ACL 2024)☆203Updated 3 months ago
- [ICML 2025] An official source code for paper "FlipAttack: Jailbreak LLMs via Flipping".☆117Updated last month
- Revisiting and Exploring Efficient Fast Adversarial Training via LAW: Lipschitz Regularization and Auto Weight Averaging (TIFS2024)☆34Updated last year
- ShieldLM: Empowering LLMs as Aligned, Customizable and Explainable Safety Detectors [EMNLP 2024 Findings]☆191Updated 8 months ago
- [NAACL 2025 Oral] 🎉 From redundancy to relevance: Enhancing explainability in multimodal large language models☆95Updated 3 months ago
- ☆131Updated 2 months ago
- Accepted by IJCAI-24 Survey Track☆205Updated 9 months ago
- This includes the original implementation of CtrlA: Adaptive Retrieval-Augmented Generation via Inherent Control.☆60Updated 7 months ago
- ☆16Updated 2 months ago
- Awesome Large Reasoning Model(LRM) Safety.This repository is used to collect security-related research on large reasoning models such as …☆64Updated this week
- Grimoire is All You Need for Enhancing Large Language Models☆115Updated last year
- RobustFT: Robust Supervised Fine-tuning for Large Language Models under Noisy Response☆41Updated 5 months ago
- Explore concepts like Self-Correct, Self-Refine, Self-Improve, Self-Contradict, Self-Play, and Self-Knowledge, alongside o1-like reasonin…☆168Updated 5 months ago
- ☆29Updated 2 months ago
- The official code of the paper "Deciphering Cross-Modal Alignment in Large Vision-Language Models with Modality Integration Rate".☆98Updated 6 months ago
- Safety at Scale: A Comprehensive Survey of Large Model Safety☆167Updated 3 months ago