☆152Aug 23, 2023Updated 2 years ago
Alternatives and similar repositories for ScanQA
Users that are interested in ScanQA are comparing it to the libraries listed below
Sorting:
- [ICLR 2023] SQA3D for embodied scene understanding and reasoning☆156Oct 13, 2023Updated 2 years ago
- [IJCAI 2022] Spatiality-guided Transformer for 3D Dense Captioning on Point Clouds (official pytorch implementation)☆21Aug 31, 2022Updated 3 years ago
- [CVPR2022 Oral] 3DJCG: A Unified Framework for Joint Dense Captioning and Visual Grounding on 3D Point Clouds☆57Jan 29, 2023Updated 3 years ago
- [ECCV2022] D3Net: A Unified Speaker-Listener Architecture for 3D Dense Captioning and Visual Grounding☆44Aug 27, 2022Updated 3 years ago
- Code for the ECCV22 paper "Bottom Up Top Down Detection Transformers for Language Grounding in Images and Point Clouds"☆95Jun 9, 2023Updated 2 years ago
- Official implementation of Language Conditioned Spatial Relation Reasoning for 3D Object Grounding (NeurIPS'22).☆66Dec 2, 2022Updated 3 years ago
- [CVPR 2021] Scan2Cap: Context-aware Dense Captioning in RGB-D Scans☆107Sep 6, 2022Updated 3 years ago
- Code accompanying our ECCV-2020 paper on 3D Neural Listeners.☆138Jun 29, 2021Updated 4 years ago
- [CVPR 2022] Multi-View Transformer for 3D Visual Grounding☆80Nov 9, 2022Updated 3 years ago
- Official implementation of ICCV 2023 paper "3D-VisTA: Pre-trained Transformer for 3D Vision and Text Alignment"☆217Sep 7, 2023Updated 2 years ago
- ☆25Mar 15, 2022Updated 3 years ago
- [ICCV 2023] Multi3DRefer: Grounding Text Description to Multiple 3D Objects☆94Oct 18, 2025Updated 4 months ago
- [TNNLS] Toward Explainable and Fine-Grained 3D Grounding through Referring Textual Phrases☆16Jul 10, 2025Updated 7 months ago
- [ECCV 2020] ScanRefer: 3D Object Localization in RGB-D Scans using Natural Language☆295Feb 10, 2023Updated 3 years ago
- Code for "Chat-3D: Data-efficiently Tuning Large Language Model for Universal Dialogue of 3D Scenes"☆56Mar 28, 2024Updated last year
- This is the code related to "Context-aware Alignment and Mutual Masking for 3D-Language Pre-training" (CVPR 2023).☆29Jun 15, 2023Updated 2 years ago
- A collection of 3D vision and language (e.g., 3D Visual Grounding, 3D Question Answering and 3D Dense Caption) papers and datasets.☆101Feb 26, 2023Updated 3 years ago
- Code for "Chat-Scene: Bridging 3D Scene and Large Language Models with Object Identifiers" (NeurIPS 2024)☆206Oct 20, 2025Updated 4 months ago
- [AAAI 24] Official Codebase for BridgeQA: Bridging the Gap between 2D and 3D Visual Question Answering: A Fusion Approach for 3D VQA☆27Jul 12, 2024Updated last year
- CLEVR3D Dataset: Comprehensive Visual Question Answering on Point Clouds through Compositional Scene Manipulation☆20Feb 2, 2024Updated 2 years ago
- Code for 3D-LLM: Injecting the 3D World into Large Language Models☆1,181Jun 6, 2024Updated last year
- [ICCV2021] 3DVG-Transformer: Relation Modeling for Visual Grounding on Point Clouds☆43Jul 6, 2022Updated 3 years ago
- ☆63May 17, 2023Updated 2 years ago
- [ECCV 2024] M3DBench introduces a comprehensive 3D instruction-following dataset with support for interleaved multi-modal prompts.☆61Oct 1, 2024Updated last year
- [ECCV 2024] Empowering 3D Visual Grounding with Reasoning Capabilities☆81Oct 10, 2024Updated last year
- Official implementation of ECCV24 paper "SceneVerse: Scaling 3D Vision-Language Learning for Grounded Scene Understanding"☆278Mar 19, 2025Updated 11 months ago
- Free-form Description-guided 3D Visual Graph Networks for Object Grounding in Point Cloud☆17Jun 23, 2022Updated 3 years ago
- [AAAI 2023 Oral] Language-Assisted 3D Feature Learning for Semantic Scene Understanding☆12Aug 1, 2023Updated 2 years ago
- [CVPR 2024 & NeurIPS 2024] EmbodiedScan: A Holistic Multi-Modal 3D Perception Suite Towards Embodied AI☆652Jun 13, 2025Updated 8 months ago
- [ICCV 2021] InstanceRefer: Cooperative Holistic Understanding for Visual Grounding on Point Clouds through Instance Multi-level Contextua…☆74Mar 22, 2025Updated 11 months ago
- [CVPR 2025] 3D-GRAND: Towards Better Grounding and Less Hallucination for 3D-LLMs☆53Jun 13, 2024Updated last year
- Awesome-LLM-3D: a curated list of Multi-modal Large Language Model in 3D world Resources☆2,117Feb 3, 2026Updated 3 weeks ago
- [CVPR 2022] X-Trans2Cap: Cross-Modal Knowledge Transfer using Transformer for 3D Dense Captioning☆36Aug 26, 2022Updated 3 years ago
- [ICCV 2025] A Simple yet Effective Pathway to Empowering LLaVA to Understand and Interact with 3D World☆373Oct 21, 2025Updated 4 months ago
- This is a PyTorch implementation of 3DRefTR proposed by our paper "A Unified Framework for 3D Point Cloud Visual Grounding"☆26Aug 24, 2023Updated 2 years ago
- [CVPR 2024] "LL3DA: Visual Interactive Instruction Tuning for Omni-3D Understanding, Reasoning, and Planning"; an interactive Large Langu…☆311Jul 17, 2024Updated last year
- ☆43Jan 17, 2024Updated 2 years ago
- SAT: 2D Semantics Assisted Training for 3D Visual Grounding, ICCV 2021 (Oral)☆33Sep 29, 2021Updated 4 years ago
- Code&Data for Grounded 3D-LLM with Referent Tokens☆132Jan 5, 2025Updated last year