InternRobotics / EgoHODLinks
Official implementation of EgoHOD at ICLR 2025; 14 EgoVis Challenge Winners in CVPR 2024
☆22Updated 6 months ago
Alternatives and similar repositories for EgoHOD
Users that are interested in EgoHOD are comparing it to the libraries listed below
Sorting:
- [ICLR 2024] Seer: Language Instructed Video Prediction with Latent Diffusion Models☆34Updated last year
- ☆49Updated 4 months ago
- ☆23Updated 3 months ago
- A curated list of Egocentric Action Understanding resources☆21Updated last month
- EgoVid-5M: A Large-Scale Video-Action Dataset for Egocentric Video Generation☆116Updated last month
- (ECCV 2024) Official repository of paper "EgoExo-Fitness: Towards Egocentric and Exocentric Full-Body Action Understanding"☆30Updated 5 months ago
- Code implementation of the paper 'FIction: 4D Future Interaction Prediction from Video'☆14Updated 6 months ago
- Official code releasse for "The Invisible EgoHand: 3D Hand Forecasting through EgoBody Pose Estimation"☆25Updated last month
- [ECCV2024, Oral, Best Paper Finalist] This is the official implementation of the paper "LEGO: Learning EGOcentric Action Frame Generation…☆37Updated 6 months ago
- Bidirectional Mapping between Action Physical-Semantic Space☆31Updated 2 weeks ago
- Official code for MotionBench (CVPR 2025)☆56Updated 6 months ago
- ☆21Updated last year
- Accepted by CVPR 2024☆38Updated last year
- ☆89Updated last month
- [ICML 2024] A Touch, Vision, and Language Dataset for Multimodal Alignment☆84Updated 3 months ago
- HandsOnVLM: Vision-Language Models for Hand-Object Interaction Prediction☆35Updated this week
- Code implementation for paper titled "HOI-Ref: Hand-Object Interaction Referral in Egocentric Vision"☆29Updated last year
- Affordance Grounding from Demonstration Video to Target Image (CVPR 2023)☆44Updated last year
- [ICLR 2025 Spotlight] Grounding Video Models to Actions through Goal Conditioned Exploration☆55Updated 4 months ago
- https://coshand.cs.columbia.edu/☆16Updated 10 months ago
- Official implementation of CVPR24 highlight paper "Move as You Say, Interact as You Can: Language-guided Human Motion Generation with Sce…☆161Updated last year
- Code and data release for the paper "Learning Object State Changes in Videos: An Open-World Perspective" (CVPR 2024)☆33Updated last year
- CVPR 2025☆30Updated 5 months ago
- Vinci: A Real-time Embodied Smart Assistant based on Egocentric Vision-Language Model☆75Updated 8 months ago
- Codes for "Affordance Diffusion: Synthesizing Hand-Object Interactions"☆128Updated 10 months ago
- [ICCV2025 Oral] Latent Motion Token as the Bridging Language for Learning Robot Manipulation from Videos☆133Updated 4 months ago
- ☆31Updated last week
- ☆17Updated last year
- Official Implementation of paper "Telling Left from Right: Identifying Geometry-Aware Semantic Correspondence"☆133Updated last month
- OST-Bench: Evaluating the Capabilities of MLLMs in Online Spatio-temporal Scene Understanding☆59Updated last month