IranQin / Awesome_World_Model_Papers
[World-Model-Survey-2024] Paper list and projects for World Model
☆9Updated 5 months ago
Alternatives and similar repositories for Awesome_World_Model_Papers:
Users that are interested in Awesome_World_Model_Papers are comparing it to the libraries listed below
- Latent Motion Token as the Bridging Language for Robot Manipulation☆81Updated 3 weeks ago
- RoboFactory: Exploring Embodied Agent Collaboration with Compositional Constraints☆38Updated this week
- Code release for "PISA Experiments: Exploring Physics Post-Training for Video Diffusion Models by Watching Stuff Drop" (arXiv 2025)☆28Updated 3 weeks ago
- Video Generation, Physical Commonsense, Semantic Adherence, VideoCon-Physics☆89Updated last month
- ☆99Updated 8 months ago
- Official repository for "iVideoGPT: Interactive VideoGPTs are Scalable World Models" (NeurIPS 2024), https://arxiv.org/abs/2405.15223☆126Updated last month
- A comprehensive list of papers investigating physical cognition in video generation, including papers, codes, and related websites.☆56Updated this week
- ☆126Updated 3 months ago
- List of papers on video-centric robot learning☆19Updated 5 months ago
- Code for FLIP: Flow-Centric Generative Planning for General-Purpose Manipulation Tasks☆58Updated 4 months ago
- ☆28Updated 4 months ago
- ☆68Updated 7 months ago
- Code for paper "Grounding Video Models to Actions through Goal Conditioned Exploration".☆44Updated 3 months ago
- ☆46Updated 4 months ago
- Single-file implementation to advance vision-language-action (VLA) models with reinforcement learning.☆50Updated this week
- MetaSpatial leverages reinforcement learning to enhance 3D spatial reasoning in vision-language models (VLMs), enabling more structured, …☆103Updated 3 weeks ago
- Official PyTorch Implementation of Unified Video Action Model (RSS 2025)☆152Updated 3 weeks ago
- EgoVid-5M: A Large-Scale Video-Action Dataset for Egocentric Video Generation☆100Updated 5 months ago
- This is the official implementation of Video Generation part of This&That: Language-Gesture Controlled Video Generation for Robot Plannin…☆38Updated 2 months ago
- The code and data of Paper: Towards World Simulator: Crafting Physical Commonsense-Based Benchmark for Video Generation☆99Updated 5 months ago
- Official implementation of "Self-Improving Video Generation"☆62Updated last month
- [ICLR2025] Official code implementation of Video-UTR: Unhackable Temporal Rewarding for Scalable Video MLLMs☆51Updated last month
- Repository for "General Flow as Foundation Affordance for Scalable Robot Learning"☆51Updated 3 months ago
- [CVPR2024] This is the official implement of MP5☆99Updated 9 months ago
- [CVPR 2024] Situational Awareness Matters in 3D Vision Language Reasoning☆37Updated 4 months ago
- [NeurIPSw'24] This repo is the official implementation of "MineDreamer: Learning to Follow Instructions via Chain-of-Imagination for Simu…☆87Updated 2 months ago
- ☆66Updated last week
- [CVPR 2025] Official implementation of ByTheWay: Boost Your Text-to-Video Generation Model to Higher Quality in a Training-free Way☆37Updated this week
- Official code for MotionBench (CVPR 2025)☆34Updated last month
- official code repo of CVPR 2025 paper PhyT2V: LLM-Guided Iterative Self-Refinement for Physics-Grounded Text-to-Video Generation☆21Updated last month