facebookresearch / IntPhys2Links
This is the code repository for IntPhys 2, a video benchmark designed to evaluate the intuitive physics understanding of deep learning models.
☆56Updated 3 weeks ago
Alternatives and similar repositories for IntPhys2
Users that are interested in IntPhys2 are comparing it to the libraries listed below
Sorting:
- This repo contains the code for the paper "Intuitive physics understanding emerges fromself-supervised pretraining on natural videos"☆169Updated 5 months ago
- Theia: Distilling Diverse Vision Foundation Models for Robot Learning☆239Updated 3 months ago
- ☆213Updated 3 months ago
- [ICML'25] The PyTorch implementation of paper: "AdaWorld: Learning Adaptable World Models with Latent Actions".☆125Updated last month
- Nvidia GEAR Lab's initiative to solve the robotics data problem using world models☆205Updated 2 weeks ago
- Cosmos-Reason1 models understand the physical common sense and generate appropriate embodied decisions in natural language through long c…☆549Updated last week
- ☆163Updated 4 months ago
- A Video Tokenizer Evaluation Dataset☆128Updated 6 months ago
- [ICLR'25] LLaRA: Supercharging Robot Learning Data for Vision-Language Policy☆214Updated 3 months ago
- OpenVLA: An open-source vision-language-action model for robotic manipulation.☆224Updated 3 months ago
- Benchmarking physical understanding in generative video models☆183Updated last month
- Cosmos-Predict1 is a collection of general-purpose world foundation models for Physical AI that can be fine-tuned into customized world m…☆295Updated last month
- Official implementation of the paper "EgoPet: Egomotion and Interaction Data from an Animal's Perspective".☆26Updated last year
- ☆120Updated 4 months ago
- Clarity: A Minimalist Website Template for AI Research☆127Updated 6 months ago
- ☆76Updated last month
- Official repository for "iVideoGPT: Interactive VideoGPTs are Scalable World Models" (NeurIPS 2024), https://arxiv.org/abs/2405.15223☆136Updated last month
- Official Reporsitory of "RoboEngine: Plug-and-Play Robot Data Augmentation with Semantic Robot Segmentation and Background Generation"☆111Updated last month
- Cosmos-Predict2 is a collection of general-purpose world foundation models for Physical AI that can be fine-tuned into customized world m…☆384Updated this week
- ☆131Updated 6 months ago
- [ICLR 2025] LAPA: Latent Action Pretraining from Videos☆326Updated 5 months ago
- Paper list in the survey: A Survey on Vision-Language-Action Models: An Action Tokenization Perspective☆110Updated 2 weeks ago
- [ICML 2024] A Touch, Vision, and Language Dataset for Multimodal Alignment☆78Updated last month
- Pytorch implementation of "Genie: Generative Interactive Environments", Bruce et al. (2024).☆166Updated 10 months ago
- NORA: A Small Open-Sourced Generalist Vision Language Action Model for Embodied Tasks☆146Updated last week
- WorldVLA: Towards Autoregressive Action World Model☆268Updated last week
- DynaMo: In-Domain Dynamics Pretraining for Visuo-Motor Control☆107Updated 8 months ago
- Unified Vision-Language-Action Model☆128Updated 2 weeks ago
- Code for subgoal synthesis via image editing☆140Updated last year
- Official repository of Learning to Act from Actionless Videos through Dense Correspondences.☆221Updated last year