rui-qian / READ
Rui Qian, Xin Yin, Dejing Dou†: Reasoning to Attend: Try to Understand How <SEG> Token Works (CVPR 2025)
☆30Updated last week
Alternatives and similar repositories for READ
Users that are interested in READ are comparing it to the libraries listed below
Sorting:
- Emerging Pixel Grounding in Large Multimodal Models Without Grounding Supervision☆41Updated last month
- [NeurIPS 2024] Official PyTorch implementation of LoTLIP: Improving Language-Image Pre-training for Long Text Understanding☆43Updated 3 months ago
- [NeurIPS 2024] Official PyTorch implementation of "Improving Compositional Reasoning of CLIP via Synthetic Vision-Language Negatives"☆39Updated 5 months ago
- The official repository for paper "PruneVid: Visual Token Pruning for Efficient Video Large Language Models".☆38Updated 2 months ago
- Repository for the paper: Teaching VLMs to Localize Specific Objects from In-context Examples☆22Updated 5 months ago
- VisualGPTScore for visio-linguistic reasoning☆27Updated last year
- [CVPR2024 Highlight] Official implementation for Transferable Visual Prompting. The paper "Exploring the Transferability of Visual Prompt…☆39Updated 4 months ago
- ☆16Updated last year
- ☆30Updated 7 months ago
- COLA: Evaluate how well your vision-language model can Compose Objects Localized with Attributes!☆24Updated 5 months ago
- (NeurIPS 2024 Spotlight) TOPA: Extend Large Language Models for Video Understanding via Text-Only Pre-Alignment☆30Updated 7 months ago
- [CVPR 2025] Mitigating Object Hallucinations in Large Vision-Language Models with Assembly of Global and Local Attention☆33Updated 9 months ago
- Official This-Is-My Dataset published in CVPR 2023☆16Updated 9 months ago
- Official Repository of Personalized Visual Instruct Tuning☆28Updated 2 months ago
- ☆21Updated 3 months ago
- ☆11Updated 10 months ago
- Offical repo for CAT-V - Caption Anything in Video: Object-centric Dense Video Captioning with Spatiotemporal Multimodal Prompting☆37Updated 2 weeks ago
- [ECCV 2024] ControlCap: Controllable Region-level Captioning☆75Updated 6 months ago
- [CVPR2025] Code Release of F-LMM: Grounding Frozen Large Multimodal Models☆86Updated 9 months ago
- [ECCV 2024] Learning Video Context as Interleaved Multimodal Sequences☆38Updated 2 months ago
- official repo for paper "[CLS] Token Tells Everything Needed for Training-free Efficient MLLMs"☆20Updated 3 weeks ago
- cliptrase☆36Updated 8 months ago
- [CVPR2024] The code of "UniPT: Universal Parallel Tuning for Transfer Learning with Efficient Parameter and Memory"☆68Updated 6 months ago
- [EMNLP 2024] Preserving Multi-Modal Capabilities of Pre-trained VLMs for Improving Vision-Linguistic Compositionality☆16Updated 7 months ago
- PyTorch code for "Contrastive Region Guidance: Improving Grounding in Vision-Language Models without Training"☆34Updated last year
- ☆24Updated 11 months ago
- NegCLIP.☆31Updated 2 years ago
- FreeVA: Offline MLLM as Training-Free Video Assistant☆61Updated 11 months ago
- [CVPR 2025 🔥]A Large Multimodal Model for Pixel-Level Visual Grounding in Videos☆64Updated 3 weeks ago
- ☆11Updated 7 months ago