Charlotte-CharMLab / FibottentionLinks
Inceptive Visual Representation Learning with Diverse Attention Across Heads. Image Classification, Action Recognition, and Robot Learning.
☆16Updated last month
Alternatives and similar repositories for Fibottention
Users that are interested in Fibottention are comparing it to the libraries listed below
Sorting:
- The offical repo for "Play to the Score: Stage-Guided Dynamic Multi-Sensory Fusion for Robotic Manipulation", CoRL 2024 (ORAL)☆12Updated 7 months ago
- Code for the paper Seeing the Pose in the Pixels: Learning Pose-Aware Representations in Vision Transformers☆21Updated 10 months ago
- Visual Relationship Reasoning for Grasp Planning☆12Updated 2 weeks ago
- LOCATE: Localize and Transfer Object Parts for Weakly Supervised Affordance Grounding (CVPR 2023)☆38Updated 2 years ago
- Winner of Cloth Competition: ICRA 2023, ICRA 2024 - Center Direction Network for Grasping Point Localization on Cloths - IEEE Robotic…☆14Updated 6 months ago
- [CVPR 2022] Joint hand motion and interaction hotspots prediction from egocentric videos☆65Updated last year
- An unofficial pytorch dataloader for Open X-Embodiment Datasets https://github.com/google-deepmind/open_x_embodiment☆14Updated 4 months ago
- Video + CLIP Baseline for Ego4D Long Term Action Anticipation Challenge (CVPR 2022)☆13Updated 2 years ago
- ☆44Updated last year
- Code for the paper: "Active Vision Might Be All You Need: Exploring Active Vision in Bimanual Robotic Manipulation"☆36Updated 2 months ago
- This repository provides the sample code designed to interpret human demonstration videos and convert them into high-level tasks for robo…☆39Updated 7 months ago
- [ICRA'24] Crossway Diffusion: Improving Diffusion-based Visuomotor Policy via Self-supervised Learning☆67Updated 10 months ago
- ☆19Updated 3 months ago
- ☆13Updated 3 months ago
- Given an RGBD image and a text prompt, ForceSight produces visual-force goals for a robot, enabling mobile manipulation in unseen environ…☆22Updated last year
- ☆12Updated last year
- Repo for Bring Your Own Vision-Language-Action (VLA) model, arxiv 2024☆28Updated 4 months ago
- This is the offical repository of LLAVIDAL☆14Updated 2 months ago
- Task-Focused Few-Shot Object Detection Benchmark☆14Updated 2 months ago
- Code & data for "RoboGround: Robotic Manipulation with Grounded Vision-Language Priors" (CVPR 2025)☆16Updated 2 weeks ago
- Code for Stable Control Representations☆25Updated 2 months ago
- [CoRL 2023] XSkill: cross embodiment skill discovery☆60Updated last year
- Code implementation for paper titled "HOI-Ref: Hand-Object Interaction Referral in Egocentric Vision"☆27Updated last year
- ☆10Updated last year
- Human Demo Videos to Robot Action Plans☆52Updated 6 months ago
- Code for "Unleashing Large-Scale Video Generative Pre-training for Visual Robot Manipulation"☆44Updated last year
- [ICRA 2025] CAGE: Causal Attention Enables Data-Efficient Generalizable Robotic Manipulation☆27Updated 4 months ago
- ☆21Updated last year
- code for the paper Predicting Point Tracks from Internet Videos enables Diverse Zero-Shot Manipulation☆87Updated 10 months ago
- ☆17Updated 6 months ago