Jingfeng0705 / LIFTLinks
The official repo for LIFT: Language-Image Alignment with Fixed Text Encoders
☆34Updated last month
Alternatives and similar repositories for LIFT
Users that are interested in LIFT are comparing it to the libraries listed below
Sorting:
- [NeurIPS 2024] TransAgent: Transfer Vision-Language Foundation Models with Heterogeneous Agent Collaboration☆24Updated 9 months ago
- ☆51Updated 6 months ago
- Official implementation of "Traceable Evidence Enhanced Visual Grounded Reasoning: Evaluation and Methodology"☆49Updated 3 weeks ago
- [Preprint] GMem: A Modular Approach for Ultra-Efficient Generative Models☆39Updated 4 months ago
- Adapting LLaMA Decoder to Vision Transformer☆29Updated last year
- Dimple, the first Discrete Diffusion Multimodal Large Language Model☆85Updated last month
- [ICLR2025] MMIU: Multimodal Multi-image Understanding for Evaluating Large Vision-Language Models☆85Updated 10 months ago
- The official implementation of the paper "MMFuser: Multimodal Multi-Layer Feature Fuser for Fine-Grained Vision-Language Understanding". …☆57Updated 9 months ago
- ZoomEye: Enhancing Multimodal LLMs with Human-Like Zooming Capabilities through Tree-Based Image Exploration☆47Updated 7 months ago
- ☆37Updated 2 months ago
- Awesome autoregressive vision foundation models☆25Updated 7 months ago
- Implementation for "The Scalability of Simplicity: Empirical Analysis of Vision-Language Learning with a Single Transformer"☆58Updated 2 weeks ago
- Official implementation of Next Block Prediction: Video Generation via Semi-Autoregressive Modeling☆38Updated 5 months ago
- [NIPS2023]Implementation of Foundation Model is Efficient Multimodal Multitask Model Selector☆37Updated last year
- ☆12Updated 6 months ago
- M2-Reasoning: Empowering MLLMs with Unified General and Spatial Reasoning☆35Updated 3 weeks ago
- X-Reasoner: Towards Generalizable Reasoning Across Modalities and Domains☆47Updated 3 months ago
- Evaluation and dataset construction code for the CVPR 2025 paper "Vision-Language Models Do Not Understand Negation"☆27Updated 3 months ago
- ☆43Updated 9 months ago
- iLLaVA: An Image is Worth Fewer Than 1/3 Input Tokens in Large Multimodal Models☆19Updated 6 months ago
- Official code for paper "GRIT: Teaching MLLMs to Think with Images"☆115Updated this week
- Fast-Slow Thinking for Large Vision-Language Model Reasoning☆17Updated 3 months ago
- [CVPR 2025] PVC: Progressive Visual Token Compression for Unified Image and Video Processing in Large Vision-Language Models☆45Updated last month
- Official Repository of Personalized Visual Instruct Tuning☆32Updated 5 months ago
- [CVPR 2024] The official implementation of paper "synthesize, diagnose, and optimize: towards fine-grained vision-language understanding"☆45Updated last month
- Official Implementation of DiffCLIP: Differential Attention Meets CLIP☆38Updated 4 months ago
- [ECCV 2024] FlexAttention for Efficient High-Resolution Vision-Language Models☆41Updated 7 months ago
- Task Preference Optimization: Improving Multimodal Large Language Models with Vision Task Alignment☆54Updated 2 weeks ago
- CLIP-MoE: Mixture of Experts for CLIP☆42Updated 9 months ago
- Official repository for LLaVA-Reward (ICCV 2025): Multimodal LLMs as Customized Reward Models for Text-to-Image Generation☆14Updated last week