jacklishufan / LaViDaLinks
Official Implementation of LaViDa: :A Large Diffusion Language Model for Multimodal Understanding
☆186Updated last week
Alternatives and similar repositories for LaViDa
Users that are interested in LaViDa are comparing it to the libraries listed below
Sorting:
- Code for MetaMorph Multimodal Understanding and Generation via Instruction Tuning☆228Updated 8 months ago
- ☆301Updated last week
- Official Implementation of Muddit [Meissonic II]: Liberating Generation Beyond Text-to-Image with a Unified Discrete Diffusion Model.☆95Updated last month
- PyTorch implementation of NEPA☆70Updated this week
- Uni-CoT: Towards Unified Chain-of-Thought Reasoning Across Text and Vision☆184Updated last week
- Machine Mental Imagery: Empower Multimodal Reasoning with Latent Visual Tokens (arXiv 2025)☆215Updated 4 months ago
- ☆79Updated 6 months ago
- Dimple, the first Discrete Diffusion Multimodal Large Language Model☆114Updated 5 months ago
- [CVPR'2025] VoCo-LLaMA: This repo is the official implementation of "VoCo-LLaMA: Towards Vision Compression with Large Language Models".☆202Updated 6 months ago
- https://huggingface.co/datasets/multimodal-reasoning-lab/Zebra-CoT☆108Updated last month
- [ICLR2025] MMIU: Multimodal Multi-image Understanding for Evaluating Large Vision-Language Models☆92Updated last year
- [NeurIPS 2025] HermesFlow: Seamlessly Closing the Gap in Multimodal Understanding and Generation☆73Updated 3 months ago
- Pixel-Level Reasoning Model trained with RL [NeuIPS25]☆256Updated last month
- TokLIP: Marry Visual Tokens to CLIP for Multimodal Comprehension and Generation☆235Updated 4 months ago
- The official code of "VL-Rethinker: Incentivizing Self-Reflection of Vision-Language Models with Reinforcement Learning" [NeurIPS25]☆169Updated 6 months ago
- An open source implementation of CLIP (With TULIP Support)☆164Updated 7 months ago
- [ICLR 2025] AuroraCap: Efficient, Performant Video Detailed Captioning and a New Benchmark☆137Updated 6 months ago
- ☆140Updated 2 months ago
- The code repository of UniRL☆47Updated 6 months ago
- ☆95Updated 6 months ago
- [COLM'25] Official implementation of the Law of Vision Representation in MLLMs☆171Updated 2 months ago
- [NeurIPS 2025] Vision as a Dialect: Unifying Visual Understanding and Generation via Text-Aligned Representations☆192Updated 3 months ago
- Code for the paper "Vamba: Understanding Hour-Long Videos with Hybrid Mamba-Transformers" [ICCV 2025]☆97Updated 4 months ago
- (ICCV 2025) "Principal Components" Enable A New Language of Images☆75Updated 4 months ago
- [arXiv: 2502.05178] QLIP: Text-Aligned Visual Tokenization Unifies Auto-Regressive Multimodal Understanding and Generation☆94Updated 9 months ago
- Official repository of 'ScaleCap: Inference-Time Scalable Image Captioning via Dual-Modality Debiasing’☆59Updated 6 months ago
- Official Code for "ARM-Thinker: Reinforcing Multimodal Generative Reward Models with Agentic Tool Use and Visual Reasoning"☆72Updated 3 weeks ago
- [CVPR2025 Highlight] PAR: Parallelized Autoregressive Visual Generation. https://yuqingwang1029.github.io/PAR-project☆184Updated 9 months ago
- Official implementation of Bifrost-1: Bridging Multimodal LLMs and Diffusion Models with Patch-level CLIP Latents (NeurIPS 2025)☆43Updated last month
- [TMLR] Public code repo for paper "A Single Transformer for Scalable Vision-Language Modeling"☆147Updated last year