ayesha-ishaq / DriveLMM-o1
Benchmark and model for step-by-step reasoning in autonomous driving.
☆35Updated last week
Alternatives and similar repositories for DriveLMM-o1:
Users that are interested in DriveLMM-o1 are comparing it to the libraries listed below
- ☆36Updated 3 weeks ago
- ☆56Updated 7 months ago
- Official PyTorch implementation of CODA-LM(https://arxiv.org/abs/2404.10595)☆86Updated 3 months ago
- Doe-1: Closed-Loop Autonomous Driving with Large World Model☆88Updated 2 months ago
- Are VLMs Ready for Autonomous Driving? An Empirical Study from the Reliability, Data, and Metric Perspectives☆59Updated last month
- Simulator designed to generate diverse driving scenarios.☆41Updated last month
- Official repository for the NuScenes-MQA. This paper is accepted by LLVA-AD Workshop at WACV 2024.☆25Updated last year
- [CVPR 2024] LaMPilot: An Open Benchmark Dataset for Autonomous Driving with Language Model Programs☆29Updated 11 months ago
- Official repository for paper "Can LVLMs Obtain a Driver’s License? A Benchmark Towards Reliable AGI for Autonomous Driving"☆28Updated last month
- [AAAI2025] Language Prompt for Autonomous Driving☆131Updated 3 months ago
- Reason2Drive: Towards Interpretable and Chain-based Reasoning for Autonomous Driving☆79Updated last year
- [NeurIPS 2024] DrivingDojo Dataset: Advancing Interactive and Knowledge-Enriched Driving World Model☆65Updated 3 months ago
- A Multi-Modal Large Language Model with Retrieval-augmented In-context Learning capacity designed for generalisable and explainable end-t…☆88Updated 5 months ago
- 【IEEE T-IV】A systematic survey of multi-modal and multi-task visual understanding foundation models for driving scenarios☆49Updated 10 months ago
- [CVPR 2024] MAPLM: A Large-Scale Vision-Language Dataset for Map and Traffic Scene Understanding☆128Updated last year
- ☆71Updated 2 weeks ago
- [ECCV 2024] Asynchronous Large Language Model Enhanced Planner for Autonomous Driving☆72Updated last month
- [Official] [IROS 2024] A goal-oriented planning to lift VLN performance for Closed-Loop Navigation: Simple, Yet Effective☆28Updated 11 months ago
- [ECCV 2024] Official implementation for "RepVF: A Unified Vector Fields Representation for Multi-task 3D Perception"☆29Updated this week
- [ICLR'25] Official Implementation of STAMP: Scalable Task And Model-agnostic Collaborative Perception☆24Updated last month
- ☆53Updated 8 months ago
- ☆22Updated 2 months ago
- Talk2BEV: Language-Enhanced Bird's Eye View Maps (ICRA'24)☆109Updated 4 months ago
- [ECCV 2024] The official code for "Dolphins: Multimodal Language Model for Driving“☆67Updated last month
- [ECCV 2024] Official GitHub repository for the paper "LingoQA: Visual Question Answering for Autonomous Driving"☆160Updated 6 months ago
- ☆61Updated 3 months ago
- [AAAI 2024] NuScenes-QA: A Multi-modal Visual Question Answering Benchmark for Autonomous Driving Scenario.☆179Updated 4 months ago
- (ICLR2025) Enhancing End-to-End Autonomous Driving with Latent World Model☆133Updated last month
- AutoTrust, a groundbreaking benchmark designed to assess the trustworthiness of DriveVLMs. This work aims to enhance public safety by ens…☆42Updated 3 months ago
- CoMamba: Real-time Cooperative Perception Unlocked with State Space Models☆21Updated 6 months ago