JulietChoo / VisionSelectorLinks
VisionSelector: End-to-End Learnable Visual Token Compression for Efficient Multimodal LLMs
☆47Updated 2 months ago
Alternatives and similar repositories for VisionSelector
Users that are interested in VisionSelector are comparing it to the libraries listed below
Sorting:
- Multi-Stage Vision Token Dropping: Towards Efficient Multimodal Large Language Model☆36Updated 11 months ago
- Code release for VTW (AAAI 2025 Oral)☆65Updated last month
- [NeurIPS'24 Oral] HydraLoRA: An Asymmetric LoRA Architecture for Efficient Fine-Tuning☆235Updated last year
- ☆124Updated last year
- ☆28Updated last year
- The official implementation for MTLoRA: A Low-Rank Adaptation Approach for Efficient Multi-Task Learning (CVPR '24)☆68Updated 5 months ago
- The official implementation of "2024NeurIPS Dynamic Tuning Towards Parameter and Inference Efficiency for ViT Adaptation"☆51Updated 11 months ago
- ☆152Updated last year
- A generalized framework for subspace tuning methods in parameter efficient fine-tuning.☆163Updated 5 months ago
- MokA: Multimodal Low-Rank Adaptation for MLLMs☆58Updated 5 months ago
- [ICLR 2025] The official pytorch implement of "Dynamic-LLaVA: Efficient Multimodal Large Language Models via Dynamic Vision-language Cont…☆66Updated 3 months ago
- [ICML'25] Official implementation of paper "SparseVLM: Visual Token Sparsification for Efficient Vision-Language Model Inference".☆210Updated last week
- Official code for paper: [CLS] Attention is All You Need for Training-Free Visual Token Pruning: Make VLM Inference Faster.☆99Updated 5 months ago
- Awesome-Low-Rank-Adaptation☆124Updated last year
- [ICLR 2025 Oral🔥] SD-LoRA: Scalable Decoupled Low-Rank Adaptation for Class Incremental Learning☆73Updated 5 months ago
- [ICCV23] Robust Mixture-of-Expert Training for Convolutional Neural Networks by Yihua Zhang, Ruisi Cai, Tianlong Chen, Guanhua Zhang, Hua…☆66Updated 2 years ago
- a training-free approach to accelerate ViTs and VLMs by pruning redundant tokens based on similarity☆40Updated 6 months ago
- [NeurIPS 2025 Spotlight] Think or Not Think: A Study of Explicit Thinking in Rule-Based Visual Reinforcement Fine-Tuning☆77Updated 3 months ago
- [CVPR 2025] PACT: Pruning and Clustering-Based Token Reduction for Faster Visual Language Models☆52Updated 2 months ago
- One-shot Entropy Minimization☆187Updated 6 months ago
- A paper list about Token Merge, Reduce, Resample, Drop for MLLMs.☆77Updated last month
- [EMNLP 2025 main 🔥] Code for "Stop Looking for Important Tokens in Multimodal Language Models: Duplication Matters More"☆95Updated 2 months ago
- ☆16Updated last year
- [ICML 2024] Model Tailor: Mitigating Catastrophic Forgetting in Multi-modal Large Language Models☆33Updated last year
- Awesome list of Mixture-of-Experts (MoE)☆24Updated last year
- State-of-the-art Parameter-Efficient MoE Fine-tuning Method☆196Updated last year
- Give us minutes, we give back a faster Mamba. The official implementation of "Faster Vision Mamba is Rebuilt in Minutes via Merged Token …☆40Updated last year
- ☆62Updated 7 months ago
- Multimodal Large Language Model (MLLM) Tuning Survey: Keeping Yourself is Important in Downstream Tuning Multimodal Large Language Model☆90Updated 4 months ago
- The official PyTorch implementation of the paper "MLAE: Masked LoRA Experts for Visual Parameter-Efficient Fine-Tuning"☆28Updated last year