inclusionAI / MingLinks
Ming - facilitating advanced multimodal understanding and generation capabilities built upon the Ling LLM.
☆558Updated last month
Alternatives and similar repositories for Ming
Users that are interested in Ming are comparing it to the libraries listed below
Sorting:
- ☆183Updated 10 months ago
- ☆285Updated 4 months ago
- ☆579Updated last month
- MiMo-VL☆611Updated 4 months ago
- ☆704Updated last month
- Official implementation of UnifiedReward & [NeurIPS 2025] UnifiedReward-Think☆651Updated last week
- ☆144Updated 4 months ago
- [NeurIPS 2025 Spotlight] A Unified Tokenizer for Visual Generation and Understanding☆491Updated last month
- The official repo for "Vidi: Large Multimodal Models for Video Understanding and Editing"☆541Updated 2 weeks ago
- Kling-Foley: Multimodal Diffusion Transformer for High-Quality Video-to-Audio Generation☆61Updated 6 months ago
- LiveCC: Learning Video LLM with Streaming Speech Transcription at Scale (CVPR 2025)☆356Updated last month
- VARGPT-v1.1: Improve Visual Autoregressive Large Unified Model via Iterative Instruction Tuning and Reinforcement Learning☆270Updated 8 months ago
- Multimodal Models in Real World☆551Updated 10 months ago
- Valley is a cutting-edge multimodal large model designed to handle a variety of tasks involving text, images, and video data.☆266Updated 3 weeks ago
- [ICML 2025] Official PyTorch implementation of LongVU☆412Updated 7 months ago
- UniWorld: High-Resolution Semantic Encoders for Unified Visual Understanding and Generation☆822Updated last month
- The official repository of the dots.vlm1 instruct models proposed by rednote-hilab.☆276Updated 3 months ago
- This is the official repo for the paper "LongCat-Flash-Omni Technical Report"☆444Updated last week
- An unified model that seamlessly integrates multimodal understanding, text-to-image generation, and image editing within a single powerfu…☆445Updated 3 weeks ago
- Official PyTorch implementation of EMOVA in CVPR 2025 (https://arxiv.org/abs/2409.18042)☆75Updated 9 months ago
- 🔥🔥First-ever hour scale video understanding models☆595Updated 5 months ago
- video-SALMONN 2 is a powerful audio-visual large language model (LLM) that generates high-quality audio-visual video captions, which is d…☆132Updated this week
- ☆80Updated 9 months ago
- Official inference code and LongText-Bench benchmark for our paper X-Omni (https://arxiv.org/pdf/2507.22058).☆398Updated 4 months ago
- AudioStory: Generating Long-Form Narrative Audio with Large Language Models☆291Updated 3 months ago
- 💡 VideoMind: A Chain-of-LoRA Agent for Long Video Reasoning☆289Updated 2 months ago
- Tarsier -- a family of large-scale video-language models, which is designed to generate high-quality video descriptions , together with g…☆507Updated 4 months ago
- ☆78Updated 7 months ago
- HumanOmni☆209Updated 9 months ago
- Fully Open Framework for Democratized Multimodal Training☆662Updated last week