arijitray1993 / SATLinks
Spatial Aptitude Training for Multimodal Langauge Models
☆20Updated last month
Alternatives and similar repositories for SAT
Users that are interested in SAT are comparing it to the libraries listed below
Sorting:
- IMProv: Inpainting-based Multimodal Prompting for Computer Vision Tasks☆58Updated last year
- ☆78Updated 6 months ago
- Code for Stable Control Representations☆26Updated 8 months ago
- HD-EPIC Python script to download the entire datasets or parts of it☆14Updated 2 months ago
- Subtask-Aware Visual Reward Learning from Segmented Demonstrations (ICLR 2025 accepted)☆18Updated 8 months ago
- [ECCV'24] 3D Reconstruction of Objects in Hands without Real World 3D Supervision☆16Updated 10 months ago
- [ICLR 2025] Official implementation and benchmark evaluation repository of <PhysBench: Benchmarking and Enhancing Vision-Language Models …☆80Updated 6 months ago
- [ICCV 2025] Official code for Perspective-Aware Reasoning in Vision-Language Models via Mental Imagery Simulation☆49Updated 3 months ago
- ☆38Updated 10 months ago
- ☆46Updated last year
- FleVRS: Towards Flexible Visual Relationship Segmentation, NeurIPS 2024☆22Updated last year
- ☆21Updated last year
- Visual Embodied Brain: Let Multimodal Large Language Models See, Think, and Control in Spaces☆87Updated 6 months ago
- LogiCity@NeurIPS'24, D&B track. A multi-agent inductive learning environment for "abstractions".☆26Updated 6 months ago
- ☆18Updated last year
- [ICCV 2023] Understanding 3D Object Interaction from a Single Image☆47Updated last year
- [CVPR 2025] 3D-GRAND: Towards Better Grounding and Less Hallucination for 3D-LLMs☆52Updated last year
- ☆42Updated 6 months ago
- [ICLR 2025 Spotlight] Grounding Video Models to Actions through Goal Conditioned Exploration☆58Updated 7 months ago
- [ICML 2024] A Touch, Vision, and Language Dataset for Multimodal Alignment☆88Updated 6 months ago
- ☆89Updated last year
- [NeurIPS 2025] Source codes for the paper "MindJourney: Test-Time Scaling with World Models for Spatial Reasoning"☆115Updated last month
- HandsOnVLM: Vision-Language Models for Hand-Object Interaction Prediction☆42Updated 3 months ago
- ☆35Updated 7 months ago
- [TMLR 2025] The official repository of the paper "Unsupervised Discovery of Object-Centric Neural Fields"☆18Updated 10 months ago
- Egocentric Video Understanding Dataset (EVUD)☆32Updated last year
- Code for paper "Super-CLEVR: A Virtual Benchmark to Diagnose Domain Robustness in Visual Reasoning"☆44Updated 2 years ago
- Official Code for the NeurIPS'23 paper "3D-Aware Visual Question Answering about Parts, Poses and Occlusions"☆19Updated last year
- Slot-TTA shows that test-time adaptation using slot-centric models can improve image segmentation on out-of-distribution examples.☆27Updated 2 years ago
- A paper list that includes world models or generative video models for embodied agents.☆25Updated 11 months ago