Uh oh!
There was an error while loading.Please reload this page.
- Notifications
You must be signed in to change notification settings - Fork78
[ECCV 2024 Oral] DriveLM: Driving with Graph Visual Question Answering
License
OpenDriveLab/DriveLM
Folders and files
| Name | Name | Last commit message | Last commit date | |
|---|---|---|---|---|
Repository files navigation
Important
🌟 Stay up to date atopendrivelab.com!
DriveLM:Driving withGraphVisualQuestionAnswering
Autonomous Driving Challenge 2024Driving-with-LanguageLeaderboard.
drivelm_nus_demo_v2_1.mp4
🔥 We instantiate datasets (DriveLM-Data) built upon nuScenes and CARLA, and propose a VLM-based baseline approach (DriveLM-Agent) for jointly performingGraph VQA and end-to-end driving.
🏁DriveLM serves as a main track in theCVPR 2024 Autonomous Driving Challenge. Everything you need for the challenge isHERE, including baseline, test data and submission format and evaluation pipeline!
[2025/01/08]Drive-Bench release! In-depth analysis in what are DriveLM really benchmarking. Take a look atarxiv.[2024/07/16]DriveLMofficial leaderboard reopen![2024/07/01]DriveLM got accepted to ECCV 2024! Congrats to the team![2024/06/01]Challenge ended up!See the final leaderboard.[2024/03/25]Challenge test server is online and the test questions are released.Check it out![2024/02/29]Challenge repo release. Baseline, data and submission format, evaluation pipeline.Have a look![2023/08/25]DriveLM-nuScenes demo released.[2023/12/22]DriveLM-nuScenes fullv1.0andpaper released.
- Highlights
- Getting Started
- Current Endeavors and Future Horizons
- TODO List
- DriveLM-Data
- License and Citation
- Other Resources
To get started with DriveLM:
- The advent of GPT-style multimodal models in real-world applications motivates the study of the role of language in driving.
- Date below reflects the arXiv submission date.
- If there is any missing work, please reach out to us!
DriveLM attempts to address some of the challenges faced by the community.
- Lack of data: DriveLM-Data serves as a comprehensive benchmark for driving with language.
- Embodiment: GVQA provides a potential direction for embodied applications of LLMs / VLMs.
- Closed-loop: DriveLM-CARLA attempts to explore closed-loop planning with language.
- DriveLM-Data
- DriveLM-nuScenes
- DriveLM-CARLA
- DriveLM-Metrics
- GPT-score
- DriveLM-Agent
- Inference code on DriveLM-nuScenes
- Inference code on DriveLM-CARLA
We facilitate thePerception, Prediction, Planning, Behavior, Motion tasks with human-written reasoning logic as a connection between them. We propose the task ofGVQA on the DriveLM-Data.
DriveLM-Data is thefirst language-driving dataset facilitating the full stack of driving tasks with graph-structured logical dependencies.
Links to details aboutGVQA task,Dataset Features, andAnnotation.
All assets and code in this repository are under theApache 2.0 license unless specified otherwise. The language data is underCC BY-NC-SA 4.0. Other datasets (including nuScenes) inherit their own distribution licenses. Please consider citing our paper and project if they help your research.
@article{sima2023drivelm,title={DriveLM: Driving with Graph Visual Question Answering},author={Sima, Chonghao and Renz, Katrin and Chitta, Kashyap and Chen, Li and Zhang, Hanxue and Xie, Chengen and Luo, Ping and Geiger, Andreas and Li, Hongyang},journal={arXiv preprint arXiv:2312.14150},year={2023}}
@misc{contributors2023drivelmrepo,title={DriveLM: Driving with Graph Visual Question Answering},author={DriveLM contributors},howpublished={\url{https://github.com/OpenDriveLab/DriveLM}},year={2023}}
OpenDriveLab
Autonomous Vision Group
About
[ECCV 2024 Oral] DriveLM: Driving with Graph Visual Question Answering
Topics
Resources
License
Code of conduct
Uh oh!
There was an error while loading.Please reload this page.
Stars
Watchers
Forks
Sponsor this project
Uh oh!
There was an error while loading.Please reload this page.
Uh oh!
There was an error while loading.Please reload this page.



