GuanxingLu / RL-BaselinesLinks
Yet another RL Baseline repo.
☆12Updated last year
Alternatives and similar repositories for RL-Baselines
Users that are interested in RL-Baselines are comparing it to the libraries listed below
Sorting:
- [arXiv 2025] MMSI-Bench: A Benchmark for Multi-Image Spatial Intelligence☆68Updated last week
- ☆116Updated 2 months ago
- [ICML 2025 Oral] Official repo of EmbodiedBench, a comprehensive benchmark designed to evaluate MLLMs as embodied agents.☆249Updated 2 months ago
- Official Implementation of CAPEAM (ICCV'23)☆16Updated last year
- Official code for "Embodied-R1: Reinforced Embodied Reasoning for General Robotic Manipulation"☆115Updated 4 months ago
- Official repository for "iVideoGPT: Interactive VideoGPTs are Scalable World Models" (NeurIPS 2024), https://arxiv.org/abs/2405.15223☆162Updated 3 months ago
- [ICCV2025] AnyBimanual: Transfering Unimanual Policy for General Bimanual Manipulation☆93Updated 6 months ago
- ☆42Updated last year
- [ICCV2025 Oral] Latent Motion Token as the Bridging Language for Learning Robot Manipulation from Videos☆158Updated 3 months ago
- ☆138Updated 6 months ago
- [ICCV 2025] RoboFactory: Exploring Embodied Agent Collaboration with Compositional Constraints☆100Updated 4 months ago
- [NeurIPS 2025] OST-Bench: Evaluating the Capabilities of MLLMs in Online Spatio-temporal Scene Understanding☆69Updated 3 months ago
- IMG: Calibrating Diffusion Models via Implicit Multimodal Guidance, ICCV 2025☆30Updated 3 months ago
- [NeurIPS 2023] MoVie: Visual Model-Based Policy Adaptation for View Generalization☆11Updated 2 years ago
- [ICLR 2025 Spotlight] Grounding Video Models to Actions through Goal Conditioned Exploration☆58Updated 8 months ago
- ICCV2025☆145Updated last month
- ☆128Updated 2 weeks ago
- [ICLR 2025] This repo is the official implementation of "The Labyrinth of Links: Navigating the Associative Maze of Multi-modal LLMs".☆13Updated 11 months ago
- Ctrl-World: A Controllable Generative World Model for Robot Manipualtion☆244Updated last month
- Data pre-processing and training code on Open-X-Embodiment with pytorch☆11Updated 11 months ago
- Being-H0: Vision-Language-Action Pretraining from Large-Scale Human Videos☆196Updated 4 months ago
- ☆47Updated last year
- VLA-RFT: Vision-Language-Action Models with Reinforcement Fine-Tuning☆113Updated 3 months ago
- ☆62Updated last year
- [CVPR2024] This is the official implement of MP5☆106Updated last year
- [AAAI26 oral] CronusVLA: Towards Efficient and Robust Manipulation via Multi-Frame Vision-Language-Action Modeling☆69Updated this week
- ☆30Updated last month
- ☆89Updated last year
- Official repository of LIBERO-plus, a generalized benchmark for in-depth robustness analysis of vision-language-action models.☆167Updated 3 weeks ago
- [ICML 2024] A Touch, Vision, and Language Dataset for Multimodal Alignment☆91Updated 7 months ago