mengcaopku / Continual-LLaVALinks
☆16Updated 8 months ago
Alternatives and similar repositories for Continual-LLaVA
Users that are interested in Continual-LLaVA are comparing it to the libraries listed below
Sorting:
- The official implementation of "2024NeurIPS Dynamic Tuning Towards Parameter and Inference Efficiency for ViT Adaptation"☆46Updated 6 months ago
- [AAAI 2025] HiRED strategically drops visual tokens in the image encoding stage to improve inference efficiency for High-Resolution Visio…☆39Updated 2 months ago
- [ICML 2024] CrossGET: Cross-Guided Ensemble of Tokens for Accelerating Vision-Language Transformers.☆33Updated 6 months ago
- Less is More: High-value Data Selection for Visual Instruction Tuning☆15Updated 5 months ago
- Instruction Tuning in Continual Learning paradigm☆53Updated 5 months ago
- This is the official repo for Debiasing Large Visual Language Models, including a Post-Hoc debias method and Visual Debias Decoding strat…☆78Updated 4 months ago
- ☆24Updated last week
- Code release for VTW (AAAI 2025) Oral☆44Updated 5 months ago
- [ICLR2025] γ -MOD: Mixture-of-Depth Adaptation for Multimodal Large Language Models☆37Updated 5 months ago
- [NeurIPS 2024] Calibrated Self-Rewarding Vision Language Models☆76Updated last year
- [ICLR 2025] VL-ICL Bench: The Devil in the Details of Multimodal In-Context Learning☆61Updated 5 months ago
- Adapting LLaMA Decoder to Vision Transformer☆28Updated last year
- [ICML 2024] Memory-Space Visual Prompting for Efficient Vision-Language Fine-Tuning☆49Updated last year
- Preventing Zero-Shot Transfer Degradation in Continual Learning of Vision-Language Models☆96Updated last year
- CorDA: Context-Oriented Decomposition Adaptation of Large Language Models for task-aware parameter-efficient fine-tuning(NeurIPS 2024)☆46Updated 6 months ago
- HalluciDoctor: Mitigating Hallucinatory Toxicity in Visual Instruction Data (Accepted by CVPR 2024)☆45Updated last year
- ☆22Updated 4 months ago
- [CVPR 2025] Mitigating Object Hallucinations in Large Vision-Language Models with Assembly of Global and Local Attention☆37Updated last year
- Implementation of "VL-Mamba: Exploring State Space Models for Multimodal Learning"☆80Updated last year
- [ICLR 2025] See What You Are Told: Visual Attention Sink in Large Multimodal Models☆34Updated 5 months ago
- HallE-Control: Controlling Object Hallucination in LMMs☆31Updated last year
- CLIP-MoE: Mixture of Experts for CLIP☆42Updated 9 months ago
- Continual Forgetting for Pre-trained Vision Models (CVPR 2024)☆65Updated this week
- Official Repository: A Comprehensive Benchmark for Logical Reasoning in MLLMs☆38Updated last month
- Github repository for "Bring Reason to Vision: Understanding Perception and Reasoning through Model Merging" (ICML 2025)☆63Updated last month
- ☆18Updated 2 months ago
- A paper list about Token Merge, Reduce, Resample, Drop for MLLMs.☆67Updated 6 months ago
- Beyond Hallucinations: Enhancing LVLMs through Hallucination-Aware Direct Preference Optimization☆89Updated last year
- [ICCV 2025] Official code for "AIM: Adaptive Inference of Multi-Modal LLMs via Token Merging and Pruning"☆31Updated 3 weeks ago
- [CVPR2024 Highlight] Official implementation for Transferable Visual Prompting. The paper "Exploring the Transferability of Visual Prompt…☆44Updated 6 months ago