Part of the book series:Lecture Notes in Computer Science ((LNCS,volume 13231))
Included in the following conference series:
2187Accesses
Abstract
Autonomous vehicle driving is gaining ground, by receiving increasing attention from the academic and industrial communities. Despite this considerable effort, there is a lack of a systematic and fair analysis of the input representations by means of a careful experimental evaluation on the same framework. To this aim, this work proposes the first comprehensive, comparative analysis of the most common inputs that can be processed by a conditional imitation learning (CIL) approach. With more details, we considered the combinations of raw and processed data—namely RGB images, depth (D) images and semantic segmentation (S)—to be assessed as inputs of the well-established Conditional Imitation Learning with ResNet and Speed prediction (CILRS) architecture. We performed a benchmark analysis, endorsed by statistical tests, on the CARLA simulator to compare the considered configurations. The achieved results showed that RGB outperformed the other monomodal inputs, in terms of success rate on the most popular benchmark NoCrash. However, RGB did not generalize well when tested on different weather conditions; overall, the best multimodal configuration was a combination of the RGB image and semantic segmentation inputs (i.e., RGBS) compared to the others, especially in regular and dense traffic scenarios. This confirms that an appropriate fusion of multimodal sensors is an effective approach in autonomous vehicle driving.
This is a preview of subscription content,log in via an institution to check access.
Access this chapter
Subscribe and save
- Get 10 units per month
- Download Article/Chapter or eBook
- 1 Unit = 1 Article or 1 Chapter
- Cancel anytime
Buy Now
- Chapter
- JPY 3498
- Price includes VAT (Japan)
- eBook
- JPY 12583
- Price includes VAT (Japan)
- Softcover Book
- JPY 15729
- Price includes VAT (Japan)
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Behl, A., Chitta, K., Prakash, A., Ohn-Bar, E., Geiger, A.: Label efficient visual abstractions for autonomous driving. In: Proceedings of IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 2338–2345. IEEE (2020)
Bojarski, M., et al.: End to end learning for self-driving cars. arXiv preprintarXiv:1604.07316 (2016)
Bojarski, M., et al.: Explaining how a deep neural network trained with end-to-end learning steers a car. arXiv preprintarXiv:1704.07911 (2017)
Chen, D., Zhou, B., Koltun, V., Krähenbühl, P.: Learning by cheating. In: Proceedings of Conference on Robot Learning, pp. 66–75. PMLR (2020)
Codevilla, F., Müller, M., López, A., Koltun, V., Dosovitskiy, A.: End-to-end driving via conditional imitation learning. In: IEEE International Conference on Robotics and Automation (ICRA), pp. 4693–4700. IEEE (2018)
Codevilla, F., Santana, E., López, A.M., Gaidon, A.: Exploring the limitations of behavior cloning for autonomous driving. In: Proceedings of IEEE/CVF International Conference on Computer Vision (ICCV), pp. 9329–9338 (2019)
Cultrera, L., Seidenari, L., Becattini, F., Pala, P., Del Bimbo, A.: Explaining autonomous driving by learning end-to-end visual attention. In: Proceedings of IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, pp. 340–341 (2020)
Deng, J., Dong, W., Socher, R., Li, L.J., Li, K., Fei-Fei, L.: ImageNet: a large-scale hierarchical image database. In: Proceedings of IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 248–255. IEEE (2009)
Dosovitskiy, A., Ros, G., Codevilla, F., Lopez, A., Koltun, V.: Carla: an open urban driving simulator. In: Conference on Robot Learning, pp. 1–16. PMLR (2017)
Eraqi, H.M., Moustafa, M.N., Honer, J.: Efficient occupancy grid mapping and camera-lidar fusion for conditional imitation learning driving. In: Proceedings of IEEE 23rd International Conference on Intelligent Transportation Systems (ITSC), pp. 1–7. IEEE (2020)
Holm, S.: A simple sequentially rejective multiple test procedure. Scand. J. Statist.6(2), 65–70 (1979).https://doi.org/10.2307/4615733
Huang, Z., Lv, C., Xing, Y., Wu, J.: Multi-modal sensor fusion-based deep neural network for end-to-end autonomous driving with scene understanding. IEEE Sensors J. MINO21(10), 11781–11790 (2020)
Ohn-Bar, E., Prakash, A., Behl, A., Chitta, K., Geiger, A.: Learning situational driving. In: Proceedings of IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 11296–11305 (2020)
Tampuu, A., Matiisen, T., Semikin, M., Fishman, D., Muhammad, N.: A survey of end-to-end driving: architectures and training methods. IEEE Trans. Neural Netw. Learn, Syst. (2020)
United States Department of Transportation: Risky Driving (2021).https://www.nhtsa.gov/. Accessed 18 Nov 2021
Wilcoxon, F.: Individual comparisons by ranking methods. Biometrics Bull.1(6), 196–202 (80–83).https://doi.org/10.2307/3001968
Xiao, Y., Codevilla, F., Gurram, A., Urfalioglu, O., López, A.M.: Multimodal end-to-end autonomous driving. IEEE Trans. Intell. Transp. Syst. (2020)
Xu, H., Gao, Y., Yu, F., Darrell, T.: End-to-end learning of driving models from large-scale video datasets. In: Proceedings of IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2174–2182 (2017)
Yu, C., Gao, C., Wang, J., Yu, G., Shen, C., Sang, N.: BiSeNet V2: Bilateral network with guided aggregation for real-time semantic segmentation. Int. J. Comput. Vis.129(11), 3051–3068 (2021)
Yurtsever, E., Lambert, J., Carballo, A., Takeda, K.: A survey of autonomous driving: common practices and emerging technologies. IEEE Access8, 58443–58469 (2020)
Author information
Authors and Affiliations
Department of Information Engineering, Electrical Engineering and Applied Mathematics, University of Salerno, 84133, Fisciano, SA, Italy
Antonio Greco, Leonardo Rundo, Alessia Saggese, Mario Vento & Antonio Vicinanza
- Antonio Greco
You can also search for this author inPubMed Google Scholar
- Leonardo Rundo
You can also search for this author inPubMed Google Scholar
- Alessia Saggese
You can also search for this author inPubMed Google Scholar
- Mario Vento
You can also search for this author inPubMed Google Scholar
- Antonio Vicinanza
You can also search for this author inPubMed Google Scholar
Corresponding author
Correspondence toAntonio Greco.
Editor information
Editors and Affiliations
Boston University, Boston, MA, USA
Stan Sclaroff
National Research Council, Lecce, Italy
Cosimo Distante
National Research Council, Lecce, Italy
Marco Leo
University of Catania, Catania, Italy
Giovanni M. Farinella
Technische Universität München, Garching, Germany
Federico Tombari
Rights and permissions
Copyright information
© 2022 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Greco, A., Rundo, L., Saggese, A., Vento, M., Vicinanza, A. (2022). Imitation Learning for Autonomous Vehicle Driving: How Does the Representation Matter?. In: Sclaroff, S., Distante, C., Leo, M., Farinella, G.M., Tombari, F. (eds) Image Analysis and Processing – ICIAP 2022. ICIAP 2022. Lecture Notes in Computer Science, vol 13231. Springer, Cham. https://doi.org/10.1007/978-3-031-06427-2_2
Download citation
Published:
Publisher Name:Springer, Cham
Print ISBN:978-3-031-06426-5
Online ISBN:978-3-031-06427-2
eBook Packages:Computer ScienceComputer Science (R0)
Share this paper
Anyone you share the following link with will be able to read this content:
Sorry, a shareable link is not currently available for this article.
Provided by the Springer Nature SharedIt content-sharing initiative