Large Vision-Language Models (LVLMs) have gained significant attention for their general knowledge and interpretability in autonomous driving. However, they lack the specialized expertise required for professional and safe driving, such as traffic rules and driving skills—critical elements of driving safety.
To address this, we introduce IDKB, a comprehensive dataset with over 1 million driving-related data items from various countries, including:
- Driving handbooks
- Theory test questions
- Simulated road test data
Similar to the process of obtaining a driver's license, IDKB encompasses knowledge from theory to practice, making it the most extensive benchmark for evaluating and enhancing LVLMs in autonomous driving.
Our benchmark includes 15 popular LVLMs evaluated for their reliability in autonomous driving scenarios, alongside fine-tuning experiments that significantly improved model performance.
- [2024-12-13] We have provided a download link via Baidu Netdisk.
- [2024-12-10] Our work, IDKB, has been accepted by AAAI 2025.
- [2024-09-04] The ArXiv version of our paper has been released.
Platform | Link | Access Code |
---|---|---|
Baidu Netdisk | Download Here | 7vfw |
Google Drive | Coming soon | - |
TODO: The usage instructions for IDKB will be provided here, stay tuned!
If you use our dataset or benchmark in your research, please cite us as:
@misc{lu2024lvlmsobtaindriverslicense,
title={Can LVLMs Obtain a Driver's License? A Benchmark Towards Reliable AGI for Autonomous Driving},
author={Yuhang Lu and Yichen Yao and Jiadong Tu and Jiangnan Shao and Yuexin Ma and Xinge Zhu},
year={2024},
eprint={2409.02914},
archivePrefix={arXiv},
primaryClass={cs.CV},
url={https://arxiv.org/abs/2409.02914},
}