NVlabs / AHALinks
A Vision-Language-Model for Detecting and Reasoning Over Failures in Robotic Manipulation
β55Updated 8 months ago
Alternatives and similar repositories for AHA
Users that are interested in AHA are comparing it to the libraries listed below
Sorting:
- β64Updated 3 months ago
- [ICLR 2025π] This is the official implementation of paper "Robots Pre-Train Robots: Manipulation-Centric Robotic Representation from Larβ¦β88Updated 11 months ago
- Code for Point Policy: Unifying Observations and Actions with Key Points for Robot Manipulationβ89Updated 5 months ago
- AutoEval: Autonomous Evaluation of Generalist Robot Manipulation Policies in the Real World | CoRL 2025β90Updated 6 months ago
- [ICRA 2025] In-Context Imitation Learning via Next-Token Predictionβ105Updated 9 months ago
- β79Updated last year
- Official Repository for SAM2Actβ216Updated 4 months ago
- A Benchmark for Low-Level Manipulation in Home Rearrangement Tasksβ168Updated 2 weeks ago
- Official Repository for MolmoActβ278Updated 2 weeks ago
- official implementation for our paper Steering Your Generalists: Improving Robotic Foundation Models via Value Guidance (CoRL 2024)β43Updated 8 months ago
- β62Updated 11 months ago
- [CoRL 25] Code for FLOWER VLA for finetuning FLOWER on CALVIN and all LIBERO environmentsβ64Updated 3 months ago
- β86Updated 3 months ago
- Code release for paper "Autonomous Improvement of Instruction Following Skills via Foundation Models" | CoRL 2024β76Updated 2 months ago
- [ICML 2025] OTTER: A Vision-Language-Action Model with Text-Aware Visual Feature Extractionβ112Updated 8 months ago
- [ICRA 2025] RACER: Rich Language-Guided Failure Recovery Policies for Imitation Learningβ40Updated last year
- Official Repository of βMotionTrans: Human VR Data Enable Motion-Level Learning for Robotic Manipulation Policies"β51Updated 3 months ago
- β53Updated 8 months ago
- Repo for Bring Your Own Vision-Language-Action (VLA) model, arxiv 2024β33Updated 11 months ago
- MOKA: Open-World Robotic Manipulation through Mark-based Visual Prompting (RSS 2024)β94Updated last year
- β75Updated last year
- Official implementation of "Towards Generalizable Vision-Language Robotic Manipulation: A Benchmark and LLM-guided 3D Policy."