showlab / Show-o
[ICLR 2025] Repository for Show-o, One Single Transformer to Unify Multimodal Understanding and Generation.
β1,369Updated last week
Alternatives and similar repositories for Show-o:
Users that are interested in Show-o are comparing it to the libraries listed below
- [CVPR 2025] The First Investigation of CoT Reasoning in Image Generationβ651Updated last month
- π This is a repository for organizing papers, codes and other resources related to unified multimodal models.β535Updated 3 weeks ago
- Autoregressive Model Beats Diffusion: π¦ Llama for Scalable Image Generationβ1,735Updated 8 months ago
- This repo contains the code for 1D tokenizer and generatorβ848Updated last month
- SEED-Voken: A Series of Powerful Visual Tokenizersβ872Updated 2 months ago
- Next-Token Prediction is All You Needβ2,106Updated last month
- [CVPR 2025 Oral]Infinity β : Scaling Bitwise AutoRegressive Modeling for High-Resolution Image Synthesisβ1,234Updated last week
- Pytorch implementation of Transfusion, "Predict the Next Token and Diffuse Images with One Multi-Modal Model", from MetaAIβ1,081Updated last month
- PyTorch implementation of MAR+DiffLoss https://arxiv.org/abs/2406.11838β1,512Updated 7 months ago
- π₯π₯π₯ A curated list of papers on LLMs-based multimodal generation (image, video, 3D and audio).β469Updated last month
- [TMLR 2025π₯] A survey for the autoregressive models in vision.β542Updated last week
- [ICLR'25 Oral] Representation Alignment for Generation: Training Diffusion Transformers Is Easier Than You Thinkβ986Updated last month
- LaVIT: Empower the Large Language Model to Understand and Generate Visual Contentβ578Updated 7 months ago
- Official PyTorch Implementation of "SiT: Exploring Flow and Diffusion-based Generative Models with Scalable Interpolant Transformers"β826Updated last year
- Liquid: Language Models are Scalable and Unified Multi-modal Generatorsβ555Updated 3 weeks ago
- This is the first paper to explore how to effectively use RL for MLLMs and introduce Vision-R1, a reasoning MLLM that leverages cold-staβ¦β540Updated 3 weeks ago
- β¨β¨[CVPR 2025] Video-MME: The First-Ever Comprehensive Evaluation Benchmark of Multi-modal LLMs in Video Analysisβ537Updated 2 weeks ago
- Explore the Multimodal βAha Momentβ on 2B Modelβ583Updated last month
- [ICLR 2025] Autoregressive Video Generation without Vector Quantizationβ488Updated 2 weeks ago
- Implementation of MagViT2 Tokenizer in Pytorchβ601Updated 3 months ago
- π₯ Sa2VA: Marrying SAM2 with LLaVA for Dense Grounded Understanding of Images and Videosβ1,070Updated last week
- Accelerating the development of large multimodal models (LMMs) with one-click evaluation module - lmms-eval.β2,416Updated this week
- [ECCV 2024] official code for "Long-CLIP: Unlocking the Long-Text Capability of CLIP"β800Updated 8 months ago
- MM-EUREKA: Exploring the Frontiers of Multimodal Reasoning with Rule-based Reinforcement Learningβ590Updated this week
- A fork to add multimodal model training to open-r1β1,245Updated 2 months ago
- [CVPR 2024] Panda-70M: Captioning 70M Videos with Multiple Cross-Modality Teachersβ601Updated 6 months ago
- [CVPR 2025 Oral] Reconstruction vs. Generation: Taming Optimization Dilemma in Latent Diffusion Modelsβ720Updated 3 weeks ago
- A family of lightweight multimodal models.β1,015Updated 5 months ago
- PyTorch implementation of RCG https://arxiv.org/abs/2312.03701β913Updated 7 months ago
- VisionLLaMA: A Unified LLaMA Backbone for Vision Tasksβ385Updated 9 months ago