MIO-Team / MIOLinks
MIO: A Foundation Model on Multimodal Tokens
☆30Updated 9 months ago
Alternatives and similar repositories for MIO
Users that are interested in MIO are comparing it to the libraries listed below
Sorting:
- UnifiedMLLM: Enabling Unified Representation for Multi-modal Multi-tasks With Large Language Model☆22Updated last year
- A project for tri-modal LLM benchmarking and instruction tuning.☆48Updated 6 months ago
- Repository for "TESS-2: A Large-Scale, Generalist Diffusion Language Model"☆50Updated 7 months ago
- Visual Programming for Text-to-Image Generation and Evaluation (NeurIPS 2023)☆56Updated 2 years ago
- [NeurIPS 2025] HermesFlow: Seamlessly Closing the Gap in Multimodal Understanding and Generation☆67Updated 3 weeks ago
- Official code implementation for the work Preference Alignment with Flow Matching (NeurIPS 2024)☆58Updated 11 months ago
- [ACL2025 Findings] Benchmarking Multihop Multimodal Internet Agents☆46Updated 7 months ago
- VPEval Codebase from Visual Programming for Text-to-Image Generation and Evaluation (NeurIPS 2023)☆44Updated last year
- AliTok: Towards Sequence Modeling Alignment between Tokenizer and Autoregressive Model☆44Updated 3 months ago
- ACDiT: Interpolating Autoregressive Conditional Modeling and Diffusion Transformer☆36Updated 9 months ago
- A big_vision inspired repo that implements a generic Auto-Encoder class capable in representation learning and generative modeling.☆34Updated last year
- The official PyTorch implementation for Improving Long-Text Alignment for Text-to-Image Diffusion Models (LongAlign)☆80Updated 5 months ago
- Explore how to get a VQ-VAE models efficiently!☆57Updated 2 months ago
- Demo page of TAVGBench: Benchmarking Text to Audible-Video Generation☆14Updated 6 months ago
- LLaVA combines with Magvit Image tokenizer, training MLLM without an Vision Encoder. Unifying image understanding and generation.☆37Updated last year
- Code and Data for Paper: SELMA: Learning and Merging Skill-Specific Text-to-Image Experts with Auto-Generated Data☆35Updated last year
- ☆18Updated 9 months ago
- [ICLR 2025] Source code for paper "A Spark of Vision-Language Intelligence: 2-Dimensional Autoregressive Transformer for Efficient Finegr…☆77Updated 10 months ago
- PyTorch implementation of StableMask (ICML'24)☆14Updated last year
- The official implementation of MAGVLT: Masked Generative Vision-and-Language Transformer (CVPR'23)☆27Updated last year
- ☆30Updated 2 months ago
- [Interspeech 2024] LiteFocus is a tool designed to accelerate diffusion-based TTA model, now implemented with the base model AudioLDM2.☆34Updated 7 months ago
- [NeurIPS 2024] Stabilize the Latent Space for Image Autoregressive Modeling: A Unified Perspective☆71Updated 11 months ago
- Pytorch Implementation of the Model from "MIRASOL3B: A MULTIMODAL AUTOREGRESSIVE MODEL FOR TIME-ALIGNED AND CONTEXTUAL MODALITIES"☆25Updated 8 months ago
- A unified framework for controllable caption generation across images, videos, and audio. Supports multi-modal inputs and customizable ca…☆51Updated 2 months ago
- ☆137Updated last year
- Towards Fine-grained Audio Captioning with Multimodal Contextual Cues☆81Updated last week
- ☆64Updated 3 months ago
- ☆78Updated 5 months ago
- [ICLR 2025] CREMA: Generalizable and Efficient Video-Language Reasoning via Multimodal Modular Fusion☆52Updated 3 months ago