XiandaGuo / Drive-MLLMLinks
[NeurIPS 2025] SURDS: Benchmarking Spatial Understanding and Reasoning in Driving Scenarios with Vision Language Models
☆66Updated 2 weeks ago
Alternatives and similar repositories for Drive-MLLM
Users that are interested in Drive-MLLM are comparing it to the libraries listed below
Sorting:
- Benchmark and model for step-by-step reasoning in autonomous driving.☆66Updated 6 months ago
- Doe-1: Closed-Loop Autonomous Driving with Large World Model☆99Updated 8 months ago
- [ICCV 2025] Are VLMs Ready for Autonomous Driving? An Empirical Study from the Reliability, Data, and Metric Perspectives☆114Updated 2 months ago
- ☆67Updated last year
- Official PyTorch implementation of CODA-LM(https://arxiv.org/abs/2404.10595)☆93Updated 10 months ago
- the official code of DriveMonkey☆34Updated 4 months ago
- [NeurIPS 2024] DrivingDojo Dataset: Advancing Interactive and Knowledge-Enriched Driving World Model☆75Updated 10 months ago
- [ECCV 2024] TOD3Cap: Towards 3D Dense Captioning in Outdoor Scenes☆129Updated 7 months ago
- ☆75Updated last month
- CoRL2024 | Hint-AD: Holistically Aligned Interpretability for End-to-End Autonomous Driving☆66Updated 11 months ago
- [CVPR 2024] MAPLM: A Large-Scale Vision-Language Dataset for Map and Traffic Scene Understanding☆152Updated last year
- [AAAI2025] Language Prompt for Autonomous Driving☆149Updated 2 weeks ago
- project page of "RAD: Training an End-to-End Driving Policy via Large-Scale 3DGS-based Reinforcement Learning"☆19Updated last week
- [ECCV 2024] Reason2Drive: Towards Interpretable and Chain-based Reasoning for Autonomous Driving☆94Updated last year
- ☆91Updated 9 months ago
- Official repository for paper "Can LVLMs Obtain a Driver’s License? A Benchmark Towards Reliable AGI for Autonomous Driving"☆29Updated 4 months ago
- [CVPR 2024] LaMPilot: An Open Benchmark Dataset for Autonomous Driving with Language Model Programs☆31Updated last year
- [ECCV 2024] The official code for "Dolphins: Multimodal Language Model for Driving“☆82Updated 7 months ago
- ☆118Updated last year
- [ECCV 2024] Make Your ViT-based Multi-view 3D Detectors Faster via Token Compression☆47Updated last year
- Official repository for the NuScenes-MQA. This paper is accepted by LLVA-AD Workshop at WACV 2024.☆31Updated last year
- ☆85Updated 10 months ago
- 【IEEE T-IV】A systematic survey of multi-modal and multi-task visual understanding foundation models for driving scenarios☆50Updated last year
- [ECCV 2024] Embodied Understanding of Driving Scenarios☆203Updated 3 months ago
- Code for CVPR2025 paper: Generating Multimodal Driving Scenes via Next-Scene Prediction☆73Updated 6 months ago
- [NeurIPS 2025] RAD: Training an End-to-End Driving Policy via Large-Scale 3DGS-based Reinforcement Learning☆48Updated this week
- Official Code Release of Delphi☆55Updated last year
- Fine-Grained Evaluation of Large Vision-Language Models in Autonomous Driving (ICCV 2025)☆25Updated 4 months ago
- [ECCV 2024] Official implementation for "RepVF: A Unified Vector Fields Representation for Multi-task 3D Perception"☆33Updated 6 months ago
- This repo contains the code for paper "LightEMMA: Lightweight End-to-End Multimodal Model for Autonomous Driving"☆111Updated last month