jingyi0000 / VLM_survey
Collection of AWESOME vision-language models for vision tasks
☆2,599Updated this week
Alternatives and similar repositories for VLM_survey:
Users that are interested in VLM_survey are comparing it to the libraries listed below
- Accelerating the development of large multimodal models (LMMs) with one-click evaluation module - lmms-eval.☆2,242Updated this week
- Project Page for "LISA: Reasoning Segmentation via Large Language Model"☆2,093Updated last month
- A collection of papers on the topic of ``Computer Vision in the Wild (CVinW)''☆1,266Updated last year
- (TPAMI 2024) A Survey on Open Vocabulary Learning☆904Updated this week
- OMG-LLaVA and OMG-Seg codebase [CVPR-24 and NeurIPS-24]☆1,262Updated 3 months ago
- Open-source evaluation toolkit of large multi-modality models (LMMs), support 220+ LMMs, 80+ benchmarks☆2,063Updated this week
- [CVPR 2024] Aligning and Prompting Everything All at Once for Universal Visual Perception☆556Updated 10 months ago
- [CVPR'23] Universal Instance Perception as Object Discovery and Retrieval☆1,266Updated last year
- [ECCV 2024] The official code of paper "Open-Vocabulary SAM".☆950Updated 7 months ago
- A family of lightweight multimodal models.☆1,006Updated 4 months ago
- Famous Vision Language Models and Their Architectures☆742Updated last month
- [CVPR 2024 🔥] Grounding Large Multimodal Model (GLaMM), the first-of-its-kind model capable of generating natural language responses tha…☆855Updated 4 months ago
- [ICLR 2025] Repository for Show-o, One Single Transformer to Unify Multimodal Understanding and Generation.☆1,284Updated this week
- [ICLR'23 Spotlight🔥] The first successful BERT/MAE-style pretraining on any convolutional network; Pytorch impl. of "Designing BERT for …☆1,335Updated last year
- 🔥 Sa2VA: Marrying SAM2 with LLaVA for Dense Grounded Understanding of Images and Videos☆991Updated last week
- This repo lists relevant papers summarized in our survey paper: A Systematic Survey of Prompt Engineering on Vision-Language Foundation …☆444Updated last week
- EVA Series: Visual Representation Fantasies from BAAI☆2,455Updated 7 months ago
- 【ICLR 2024🔥】 Extending Video-Language Pretraining to N-modality by Language-based Semantic Alignment☆795Updated last year
- VisionLLM Series☆1,031Updated last month
- Recent LLM-based CV and related works. Welcome to comment/contribute!☆859Updated 2 weeks ago
- Eagle Family: Exploring Model Designs, Data Recipes and Training Strategies for Frontier-Class Multimodal LLMs☆635Updated 2 months ago
- ☆503Updated 4 months ago
- [CVPR 2024] Alpha-CLIP: A CLIP Model Focusing on Wherever You Want☆793Updated 7 months ago
- Code for ALBEF: a new vision-language pre-training method☆1,622Updated 2 years ago
- [CVPR2024 Highlight]GLEE: General Object Foundation Model for Images and Videos at Scale☆1,111Updated 5 months ago
- 【CVPR 2024 Highlight】Monkey (LMM): Image Resolution and Text Label Are Important Things for Large Multi-modal Models☆1,735Updated last week
- Grounded Language-Image Pre-training☆2,362Updated last year
- Awesome list for research on CLIP (Contrastive Language-Image Pre-Training).☆1,183Updated 9 months ago
- Official repository of ’Visual-RFT: Visual Reinforcement Fine-Tuning’☆1,384Updated last week
- A Framework of Small-scale Large Multimodal Models☆778Updated this week