Interactive Attention Learning on Detection of Lane and Lane Marking on the Road by Monocular Camera Image
Author:
Tian Wei1ORCID, Yu Xianwang1, Hu Haohao2ORCID
Affiliation:
1. Tongji University, Shanghai 201804, China 2. Institute of Measurement and Control Systems, Karlsruhe Institute of Technology, 76131 Karlsruhe, Germany
Abstract
Vision-based identification of lane area and lane marking on the road is an indispensable function for intelligent driving vehicles, especially for localization, mapping and planning tasks. However, due to the increasing complexity of traffic scenes, such as occlusion and discontinuity, detecting lanes and lane markings from an image captured by a monocular camera becomes persistently challenging. The lanes and lane markings have a strong position correlation and are constrained by a spatial geometry prior to the driving scene. Most existing studies only explore a single task, i.e., either lane marking or lane detection, and do not consider the inherent connection or exploit the modeling of this kind of relationship between both elements to improve the detection performance of both tasks. In this paper, we establish a novel multi-task encoder–decoder framework for the simultaneous detection of lanes and lane markings. This approach deploys a dual-branch architecture to extract image information from different scales. By revealing the spatial constraints between lanes and lane markings, we propose an interactive attention learning for their feature information, which involves a Deformable Feature Fusion module for feature encoding, a Cross-Context module as information decoder, a Cross-IoU loss and a Focal-style loss weighting for robust training. Without bells and whistles, our method achieves state-of-the-art results on tasks of lane marking detection (with 32.53% on IoU, 81.61% on accuracy) and lane segmentation (with 91.72% on mIoU) of the BDD100K dataset, which showcases an improvement of 6.33% on IoU, 11.11% on accuracy in lane marking detection and 0.22% on mIoU in lane detection compared to the previous methods.
Funder
National Natural Science Foundation of China Shanghai Science and Technology Commission original research project of Tongji University National Key R&D Program of China Natural Science Foundation of Chongqing
Subject
Electrical and Electronic Engineering,Biochemistry,Instrumentation,Atomic and Molecular Physics, and Optics,Analytical Chemistry
Reference56 articles.
1. HERE (2022, December 01). HERE HD Live Map: The Most Intelligent Sensor for Autonomous Driving. Available online: https://www.here.com/platform/automotive-services/hd-maps. 2. TomTom (2022, December 01). HD Maps—Highly Accurate Border-to-Border Model of the Road. Available online: https://www.tomtom.com/products/hd-map. 3. Homayounfar, N., Ma, W.C., Liang, J., Wu, X., Fan, J., and Urtasun, R. (2019, January 15–20). DAGmapper: Learning to map by discovering lane topology. Proceedings of the IEEE/CVF International Conference on Computer Vision (CVPR), Long Beach, CA, USA. 4. Chiu, K.Y., and Lin, S.F. (2005, January 6–8). Lane detection using color-based segmentation. Proceedings of the IEEE Intelligent Vehicles Symposium, Las Vegas, NV, USA. 5. Hierarchical additive Hough transform for lane detection;Satzoda;IEEE Embed. Syst. Lett.,2010
Cited by
1 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
|
|