cofe-ai / O2-MAGVIT2Links
Open Source Implementation of Dual Modality MAGVIT2 Tokenizer
☆21Updated 9 months ago
Alternatives and similar repositories for O2-MAGVIT2
Users that are interested in O2-MAGVIT2 are comparing it to the libraries listed below
Sorting:
- SEED-Voken: A Series of Powerful Visual Tokenizers☆934Updated 2 months ago
- Implementation of MagViT2 Tokenizer in Pytorch☆629Updated 7 months ago
- Pytorch implementation of Transfusion, "Predict the Next Token and Diffuse Images with One Multi-Modal Model", from MetaAI☆1,201Updated 2 months ago
- This repo contains the code for 1D tokenizer and generator☆1,017Updated 5 months ago
- A flexible and efficient codebase for training visually-conditioned language models (VLMs)☆785Updated last year
- [ICLR 2025] VILA-U: a Unified Foundation Model Integrating Visual Understanding and Generation☆383Updated 4 months ago
- [CVPR 2025] 🔥 Official impl. of "TokenFlow: Unified Image Tokenizer for Multimodal Understanding and Generation".☆374Updated last month
- Long-RL: Scaling RL to Long Sequences☆603Updated 2 weeks ago
- PyTorch implementation of MAR+DiffLoss https://arxiv.org/abs/2406.11838☆1,724Updated 11 months ago
- LaVIT: Empower the Large Language Model to Understand and Generate Visual Content☆590Updated 11 months ago
- [Survey] Next Token Prediction Towards Multimodal Intelligence: A Comprehensive Survey☆448Updated 7 months ago
- Official PyTorch Implementation of "SiT: Exploring Flow and Diffusion-based Generative Models with Scalable Interpolant Transformers"☆959Updated last year
- [ECCV 2024 Oral] Code for paper: An Image is Worth 1/2 Tokens After Layer 2: Plug-and-Play Inference Acceleration for Large Vision-Langua…☆481Updated 8 months ago
- EVE Series: Encoder-Free Vision-Language Models from BAAI☆349Updated last month
- [ICLR 2025] Repository for Show-o series, One Single Transformer to Unify Multimodal Understanding and Generation.☆1,689Updated this week
- Resources and paper list for "Thinking with Images for LVLMs". This repository accompanies our survey on how LVLMs can leverage visual in…☆911Updated last week
- code for "Diffusion Forcing: Next-token Prediction Meets Full-Sequence Diffusion"☆1,004Updated 5 months ago
- Official Repo for Fine-Tuning Large Vision-Language Models as Decision-Making Agents via Reinforcement Learning☆384Updated 8 months ago
- High-performance Image Tokenizers for VAR and AR☆286Updated 4 months ago
- An open-source implementaion for fine-tuning Qwen2-VL and Qwen2.5-VL series by Alibaba Cloud.☆1,125Updated last week
- [AAAI-25] Cobra: Extending Mamba to Multi-modal Large Language Model for Efficient Inference☆287Updated 8 months ago
- 📖 This is a repository for organizing papers, codes and other resources related to unified multimodal models.☆678Updated last month
- A collection of resources and papers on Vector Quantized Variational Autoencoder (VQ-VAE) and its application☆306Updated 7 months ago
- Video-R1: Reinforcing Video Reasoning in MLLMs [🔥the first paper to explore R1 for video]☆676Updated last month
- [ECCV 2024] Official PyTorch implementation of RoPE-ViT "Rotary Position Embedding for Vision Transformer"☆393Updated 8 months ago
- [ICLR'25 Oral] Representation Alignment for Generation: Training Diffusion Transformers Is Easier Than You Think☆1,293Updated 5 months ago
- MM-Eureka V0 also called R1-Multimodal-Journey, Latest version is in MM-Eureka☆317Updated 2 months ago
- ⭐️ Reason-RFT: Reinforcement Fine-Tuning for Visual Reasoning.☆197Updated last month
- Visualizing the attention of vision-language models☆229Updated 6 months ago
- [TMLR 2025🔥] A survey for the autoregressive models in vision.☆691Updated this week