nvidia-cosmos / cosmos-rlLinks
Cosmos-RL is a flexible and scalable Reinforcement Learning framework specialized for Physical AI applications.
☆138Updated this week
Alternatives and similar repositories for cosmos-rl
Users that are interested in cosmos-rl are comparing it to the libraries listed below
Sorting:
- ☆23Updated last month
- EO: Open-source Unified Embodied Foundation Model Series☆29Updated this week
- Multi-SpatialMLLM Multi-Frame Spatial Understanding with Multi-Modal Large Language Models☆147Updated 3 months ago
- Code release for paper "Test-Time Training Done Right"☆275Updated this week
- A curated list of recent papers on efficient video attention for video diffusion models, including sparsification, quantization, and cach…☆36Updated last month
- Official implementation for BitVLA: 1-bit Vision-Language-Action Models for Robotics Manipulation☆73Updated last month
- Cosmos-Predict2 is a collection of general-purpose world foundation models for Physical AI that can be fine-tuned into customized world m…☆527Updated this week
- Cosmos-Predict1 is a collection of general-purpose world foundation models for Physical AI that can be fine-tuned into customized world m…☆331Updated 2 weeks ago
- Official repository for "RLVR-World: Training World Models with Reinforcement Learning", https://arxiv.org/abs/2505.13934☆80Updated 2 months ago
- A Video Tokenizer Evaluation Dataset☆132Updated 7 months ago
- Long-RL: Scaling RL to Long Sequences☆597Updated 2 weeks ago
- [CVPR 2025 Highlight] Towards Autonomous Micromobility through Scalable Urban Simulation☆106Updated last month
- ☆136Updated 7 months ago
- RLinf is a flexible and scalable open-source infrastructure designed for post-training foundation models (LLMs, VLMs, VLAs) via reinforce…☆49Updated last week
- ☆240Updated 2 months ago
- [ICLR 2025] Official implementation and benchmark evaluation repository of <PhysBench: Benchmarking and Enhancing Vision-Language Models …☆68Updated 3 months ago
- Virtual Community: An Open World for Humans, Robots, and Society☆169Updated last week
- Source codes for the paper "MindJourney: Test-Time Scaling with World Models for Spatial Reasoning"☆79Updated last month
- [ICML'25] The PyTorch implementation of paper: "AdaWorld: Learning Adaptable World Models with Latent Actions".☆147Updated 2 months ago
- Nvidia GEAR Lab's initiative to solve the robotics data problem using world models☆289Updated last week
- Visual Embodied Brain: Let Multimodal Large Language Models See, Think, and Control in Spaces☆80Updated 2 months ago
- Memory Efficient Training Framework for Large Video Generation Model☆25Updated last year
- Code for Draft Attention☆90Updated 3 months ago
- ☆144Updated 2 weeks ago
- Official implementation of CEED-VLA: Consistency Vision-Language-Action Model with Early-Exit Decoding.☆29Updated last month
- Visual Planning: Let's Think Only with Images☆269Updated 3 months ago
- Cosmos-Reason1 models understand the physical common sense and generate appropriate embodied decisions in natural language through long c…☆664Updated last week
- Efficient triton implementation of Native Sparse Attention.☆209Updated 3 months ago
- An ML research template with good documentation by Boyuan Chen, an MIT PhD student☆80Updated 6 months ago
- EgoVid-5M: A Large-Scale Video-Action Dataset for Egocentric Video Generation☆114Updated last month