TtuHamg / TextToucherLinks
Official Pytorch Implementation for "TextToucher: Fine-Grained Text-to-Touch Generation" (AAAI 2025)
☆17Updated 8 months ago
Alternatives and similar repositories for TextToucher
Users that are interested in TextToucher are comparing it to the libraries listed below
Sorting:
- [CVPR 2024] Binding Touch to Everything: Learning Unified Multimodal Tactile Representations☆73Updated last month
- ☆128Updated 2 weeks ago
- Official implemetation of the paper "InSpire: Vision-Language-Action Models with Intrinsic Spatial Reasoning"☆47Updated last month
- An example RLDS dataset builder for X-embodiment dataset conversion.☆55Updated 10 months ago
- [ICML 2024] A Touch, Vision, and Language Dataset for Multimodal Alignment☆91Updated 7 months ago
- [ICML 2025] OTTER: A Vision-Language-Action Model with Text-Aware Visual Feature Extraction☆112Updated 8 months ago
- A collection of paper/projects that trains flow matching model/policies via RL.☆342Updated 2 weeks ago
- [ICCV2025] AnyBimanual: Transfering Unimanual Policy for General Bimanual Manipulation☆93Updated 6 months ago
- An unofficial pytorch dataloader for Open X-Embodiment Datasets https://github.com/google-deepmind/open_x_embodiment☆23Updated last year
- Official PyTorch Implementation of Learning Affordance Grounding from Exocentric Images, CVPR 2022☆71Updated last year
- [ICLR 2025] LAPA: Latent Action Pretraining from Videos☆434Updated 11 months ago
- [AAAI 2026] Official code for MoLe-VLA: Dynamic Layer-skipping Vision Language Action Model via Mixture-of-Layers for Efficient Robot Man…☆60Updated 5 months ago
- [NeurIPS 24] Incremental Learning of Retrievable Skills For Efficient Continual Task Adaptation☆20Updated 3 months ago
- ICCV2025☆145Updated last month
- official repo for AGNOSTOS, a cross-task manipulation benchmark, and X-ICM method, a cross-task in-context manipulation (VLA) method☆53Updated last month
- [NeurIPS 2025] VLA-Cache: Towards Efficient Vision-Language-Action Model via Adaptive Token Caching in Robotic Manipulation☆58Updated 3 months ago
- Official repository of Learning to Act from Actionless Videos through Dense Correspondences.☆245Updated last year
- [ICCV2025 Oral] Latent Motion Token as the Bridging Language for Learning Robot Manipulation from Videos☆158Updated 3 months ago
- [ICLR 2024] Seer: Language Instructed Video Prediction with Latent Diffusion Models☆34Updated last year
- Official repo of VLABench, a large scale benchmark designed for fairly evaluating VLA, Embodied Agent, and VLMs.☆363Updated 2 months ago
- GRAPE: Guided-Reinforced Vision-Language-Action Preference Optimization☆154Updated 9 months ago
- Official repository for "iVideoGPT: Interactive VideoGPTs are Scalable World Models" (NeurIPS 2024), https://arxiv.org/abs/2405.15223☆162Updated 3 months ago
- ☆12Updated 8 months ago
- ☆423Updated 3 weeks ago
- An official implementation of Touch100k: A Large-Scale Touch-Language-Vision Dataset for Touch-Centric Multimodal Representation☆31Updated last year
- Code of "MemoryVLA: Perceptual-Cognitive Memory in Vision-Language-Action Models for Robotic Manipulation"☆114Updated last month
- [RA-L 2025] Motion Before Action: Diffusing Object Motion as Manipulation Condition☆67Updated 2 months ago
- Official PyTorch Implementation of Unified Video Action Model (RSS 2025)☆308Updated 5 months ago
- Preview code of ECCV'24 paper "Distill Gold from Massive Ores" (BiLP)☆25Updated last year
- https://arxiv.org/pdf/2506.06677☆43Updated 2 months ago