Wu-Zongyu / LanP
Official Implementation of 'Lanp: Rethinking the Impact of Language Priors in Large Vision-Language Models'
☆13Updated 2 months ago
Alternatives and similar repositories for LanP:
Users that are interested in LanP are comparing it to the libraries listed below
- ☆47Updated 5 months ago
- OOD Generalization相关文章的阅读笔记☆31Updated 4 months ago
- Awesome-Efficient-Inference-for-LRMs is a collection of state-of-the-art, novel, exciting, token-efficient methods for Large Reasoning Mo…☆59Updated last week
- ☆27Updated this week
- A paper list about large language models and multimodal models (Diffusion, VLM). From foundations to applications. It is only used to rec…☆55Updated 2 months ago
- [NeurIPS 2024] Fight Back Against Jailbreaking via Prompt Adversarial Tuning☆10Updated 5 months ago
- [NAACL 25 Demo] TrustEval: A modular and extensible toolkit for comprehensive trust evaluation of generative foundation models (GenFMs)☆97Updated 3 weeks ago
- ☆34Updated 2 months ago
- Awesome Large Reasoning Model(LRM) Safety.This repository is used to collect security-related research on large reasoning models such as …☆63Updated this week
- AlphaEdit: Null-Space Constrained Knowledge Editing for Language Models, ICLR 2025 (Outstanding Paper)☆167Updated this week
- A survey on harmful fine-tuning attack for large language model☆161Updated last week
- [ECCV 2024] The official code for "AdaShield: Safeguarding Multimodal Large Language Models from Structure-based Attack via Adaptive Shi…☆57Updated 9 months ago
- A tiny paper rating web☆36Updated last month
- ☆33Updated 2 weeks ago
- [ICLR 2025] BlueSuffix: Reinforced Blue Teaming for Vision-Language Models Against Jailbreak Attacks☆17Updated 2 weeks ago
- [NeurIPS 2024] "Can Language Models Perform Robust Reasoning in Chain-of-thought Prompting with Noisy Rationales?"☆35Updated 3 months ago
- Official Repository for The Paper: Safety Alignment Should Be Made More Than Just a Few Tokens Deep☆98Updated this week
- 📜 Paper list on decoding methods for LLMs and LVLMs☆39Updated 3 months ago
- Data and Code for Paper VLSBench: Unveiling Visual Leakage in Multimodal Safety☆37Updated last month
- This is the official code for the paper "Safety Tax: Safety Alignment Makes Your Large Reasoning Models Less Reasonable".☆14Updated last month
- The official implementation of ECCV'24 paper "To Generate or Not? Safety-Driven Unlearned Diffusion Models Are Still Easy To Generate Uns…☆73Updated 2 months ago
- ☆20Updated this week
- An implementation for MLLM oversensitivity evaluation☆13Updated 5 months ago
- This repo is for the safety topic, including attacks, defenses and studies related to reasoning and RL☆17Updated this week
- ☆21Updated last month
- "In-Context Unlearning: Language Models as Few Shot Unlearners". Martin Pawelczyk, Seth Neel* and Himabindu Lakkaraju*; ICML 2024.☆25Updated last year
- Survey on Data-centric Large Language Models☆83Updated 9 months ago
- The First to Know: How Token Distributions Reveal Hidden Knowledge in Large Vision-Language Models?☆28Updated 5 months ago
- V1: Toward Multimodal Reasoning by Designing Auxiliary Task☆33Updated 2 weeks ago
- This is the official code for the paper "Booster: Tackling Harmful Fine-tuning for Large Language Models via Attenuating Harmful Perturba…☆25Updated last month