huiwon-jang / RSPLinks
Visual Representation Learning with Stochastic Frame Prediction (ICML 2024)
☆22Updated 7 months ago
Alternatives and similar repositories for RSP
Users that are interested in RSP are comparing it to the libraries listed below
Sorting:
- Code release for ICLR 2023 paper: SlotFormer on object-centric dynamics models☆112Updated last year
- Official Code for Neural Systematic Binder☆33Updated 2 years ago
- ☆56Updated 2 years ago
- Official code for Slot-Transformer for Videos (STEVE)☆48Updated 2 years ago
- Code for Stable Control Representations☆25Updated 3 months ago
- VP2 Benchmark (A Control-Centric Benchmark for Video Prediction, ICLR 2023)☆27Updated 4 months ago
- ☆42Updated last year
- Masked World Models for Visual Control☆126Updated 2 years ago
- PyTorch implementation of the Hiveformer research paper☆48Updated 2 years ago
- Official repository for "LIV: Language-Image Representations and Rewards for Robotic Control" (ICML 2023)☆110Updated last year
- [WIP] Code for LangToMo☆14Updated 3 weeks ago
- ☆44Updated last year
- The repository of ICML2023 paper: On Pre-Training for Visuo-Motor Control: Revisiting a Learning-from-Scratch Baseline☆23Updated 2 years ago
- ☆79Updated 2 years ago
- ☆42Updated last year
- This repository is the official implementation of Improving Object-centric Learning With Query Optimization☆50Updated 2 years ago
- ☆11Updated last year
- ☆12Updated last year
- [NeurIPS 2024] GenRL: Multimodal-foundation world models enable grounding language and video prompts into embodied domains, by turning th…☆78Updated 3 months ago
- [ICML 2025] OTTER: A Vision-Language-Action Model with Text-Aware Visual Feature Extraction☆88Updated 3 months ago
- IMProv: Inpainting-based Multimodal Prompting for Computer Vision Tasks☆57Updated 9 months ago
- NeurIPS 2022 Paper "VLMbench: A Compositional Benchmark for Vision-and-Language Manipulation"☆93Updated 2 months ago
- [ICRA2023] Grounding Language with Visual Affordances over Unstructured Data☆43Updated last year
- The repository for a thorough empirical evaluation of pre-trained vision model performance across different downstream policy learning me…☆23Updated last year
- Code for the ICLR 2024 spotlight paper: "Learning to Act without Actions" (introducing Latent Action Policies)☆116Updated 11 months ago
- Official repository for "VIP: Towards Universal Visual Reward and Representation via Value-Implicit Pre-Training"☆162Updated last year
- ☆12Updated 2 months ago
- [ICLR 2025 Spotlight] Grounding Video Models to Actions through Goal Conditioned Exploration☆49Updated 2 months ago
- Official codebase for EmbCLIP☆126Updated 2 years ago
- ☆75Updated 10 months ago