zh460045050 / V2L-TokenizerLinks
☆137Updated last year
Alternatives and similar repositories for V2L-Tokenizer
Users that are interested in V2L-Tokenizer are comparing it to the libraries listed below
Sorting:
- [ECCV 2024] Official PyTorch implementation of DreamLIP: Language-Image Pre-training with Long Captions☆136Updated 4 months ago
- ☆91Updated 2 years ago
- [CVPR2025] Code Release of F-LMM: Grounding Frozen Large Multimodal Models☆103Updated 3 months ago
- ☆118Updated last year
- [NeurIPS 2024] Visual Perception by Large Language Model’s Weights☆45Updated 5 months ago
- [CVPR 2025] RAP: Retrieval-Augmented Personalization☆69Updated last month
- CLAP: Isolating Content from Style through Contrastive Learning with Augmented Prompts☆54Updated last year
- [ICCV 2023 Oral] Official Implementation of "Denoising Diffusion Autoencoders are Unified Self-supervised Learners"☆178Updated last year
- A curated list of publications on image and video segmentation leveraging Multimodal Large Language Models (MLLMs), highlighting state-of…☆120Updated last week
- [ICCV 2023] Generative Prompt Model for Weakly Supervised Object Localization☆57Updated last year
- [NeurIPS 2024] Classification Done Right for Vision-Language Pre-Training☆216Updated 5 months ago
- High-performance Image Tokenizers for VAR and AR☆286Updated 4 months ago
- DenseFusion-1M: Merging Vision Experts for Comprehensive Multimodal Perception☆154Updated 9 months ago
- [ICLR 2024 (Spotlight)] "Frozen Transformers in Language Models are Effective Visual Encoder Layers"☆243Updated last year
- [ICLR 2025] Diffusion Feedback Helps CLIP See Better☆289Updated 7 months ago
- LLaVA-NeXT-Image-Llama3-Lora, Modified from https://github.com/arielnlee/LLaVA-1.6-ft☆44Updated last year
- ☆74Updated last month
- Think or Not Think: A Study of Explicit Thinking in Rule-Based Visual Reinforcement Fine-Tuning☆62Updated 3 months ago
- ☆113Updated last year
- [ICCV 2023] ALIP: Adaptive Language-Image Pre-training with Synthetic Caption☆98Updated last year
- [ICLR2025] This repository is the official implementation of our Autoregressive Pretraining with Mamba in Vision☆84Updated 3 months ago
- HiMTok: Learning Hierarchical Mask Tokens for Image Segmentation with Large Multimodal Model☆65Updated last month
- [ICLR'25] Reconstructive Visual Instruction Tuning☆114Updated 5 months ago
- ☆71Updated 4 months ago
- [ICLR2024 Spotlight] Code Release of CLIPSelf: Vision Transformer Distills Itself for Open-Vocabulary Dense Prediction☆193Updated last year
- Official PyTorch implementation for "Diffusion Models and Semi-Supervised Learners Benefit Mutually with Few Labels"☆95Updated last year
- Visual self-questioning for large vision-language assistant.☆43Updated last month
- [CVPR 2024] Official implementation of "ViTamin: Designing Scalable Vision Models in the Vision-language Era"☆209Updated last year
- [ECCV 2024] ControlCap: Controllable Region-level Captioning☆79Updated 10 months ago
- [CVPR 2025] Mono-InternVL: Pushing the Boundaries of Monolithic Multimodal Large Language Models with Endogenous Visual Pre-training☆81Updated last month