aim-uofa / Omni-R1Links
Official Repo of Omni-R1: Reinforcement Learning for Omnimodal Reasoning via Two-System Collaboration
☆63Updated 3 weeks ago
Alternatives and similar repositories for Omni-R1
Users that are interested in Omni-R1 are comparing it to the libraries listed below
Sorting:
- ACTIVE-O3: Empowering Multimodal Large Language Models with Active Perception via GRPO☆59Updated 3 weeks ago
- Code release for "PISA Experiments: Exploring Physics Post-Training for Video Diffusion Models by Watching Stuff Drop" (ICML 2025)☆35Updated last month
- A list of works on video generation towards world model☆151Updated this week
- [CVPR 2025] The code for paper ''Video-3D LLM: Learning Position-Aware Video Representation for 3D Scene Understanding''.☆117Updated 3 weeks ago
- [CVPR 2025 (Oral)] Open implementation of "RandAR"☆175Updated 3 months ago
- SpaceR: The first MLLM empowered by SG-RLVR for video spatial reasoning☆63Updated 2 weeks ago
- ☆30Updated 6 months ago
- Official implementation of Spatial-MLLM: Boosting MLLM Capabilities in Visual-based Spatial Intelligence☆235Updated this week
- WISE: A World Knowledge-Informed Semantic Evaluation for Text-to-Image Generation☆120Updated 2 weeks ago
- [ICCV2025]Code Release of Harmonizing Visual Representations for Unified Multimodal Understanding and Generation☆126Updated last month
- Unified Vision-Language-Action Model☆61Updated this week
- A comprehensive list of papers investigating physical cognition in video generation, including papers, codes, and related websites.☆121Updated last week
- Official repo for "GigaTok: Scaling Visual Tokenizers to 3 Billion Parameters for Autoregressive Image Generation"☆162Updated 2 months ago
- GenDoP: Auto-regressive Camera Trajectory Generation as a Director of Photography☆64Updated 3 weeks ago
- [arXiv: 2502.05178] QLIP: Text-Aligned Visual Tokenization Unifies Auto-Regressive Multimodal Understanding and Generation☆75Updated 3 months ago
- [ICML2025] The code and data of Paper: Towards World Simulator: Crafting Physical Commonsense-Based Benchmark for Video Generation☆112Updated 8 months ago
- Code for MetaMorph Multimodal Understanding and Generation via Instruction Tuning☆191Updated 2 months ago
- [CVPR 25] A framework named B^2-DiffuRL for RL-based diffusion model fine-tuning.☆30Updated 2 months ago
- [CVPR 2025] Science-T2I: Addressing Scientific Illusions in Image Synthesis☆56Updated last month
- A collection of vision foundation models unifying understanding and generation.☆55Updated 5 months ago
- MetaSpatial leverages reinforcement learning to enhance 3D spatial reasoning in vision-language models (VLMs), enabling more structured, …☆133Updated last month
- Official Implementation of LaViDa: :A Large Diffusion Language Model for Multimodal Understanding☆96Updated last week
- Multi-SpatialMLLM Multi-Frame Spatial Understanding with Multi-Modal Large Language Models☆123Updated last month
- Video-Holmes: Can MLLM Think Like Holmes for Complex Video Reasoning?☆51Updated 3 weeks ago
- Official repository of 'ScaleCap: Inference-Time Scalable Image Captioning via Dual-Modality Debiasing’☆33Updated this week
- [ARXIV’25] Learning Video Generation for Robotic Manipulation with Collaborative Trajectory Control☆64Updated 3 weeks ago
- PyTorch implementation of DiffMoE, TC-DiT, EC-DiT and Dense DiT☆113Updated 2 months ago
- [arXiv 2025] MMSI-Bench: A Benchmark for Multi-Image Spatial Intelligence☆37Updated last week
- Video Generation, Physical Commonsense, Semantic Adherence, VideoCon-Physics☆119Updated last month
- [CVPR 2024] Customize your NeRF: Adaptive Source Driven 3D Scene Editing via Local-Global Iterative Training☆40Updated last year