Multi-Task Environmental Perception Methods for Autonomous Driving
Author:
Liu Ri1, Yang Shubin1, Tang Wansha1, Yuan Jie1, Chan Qiqing1, Yang Yunchuan1
Affiliation:
1. School of Electrical and Information Engineering, Wuhan Institute of Technology, Wuhan 430205, China
Abstract
In autonomous driving, environmental perception technology often encounters challenges such as false positives, missed detections, and low accuracy, particularly in detecting small objects and complex scenarios. Existing algorithms frequently suffer from issues like feature redundancy, insufficient contextual interaction, and inadequate information fusion, making it difficult to perform multi-task detection and segmentation efficiently. To address these challenges, this paper proposes an end-to-end multi-task environmental perception model named YOLO-Mg, designed to simultaneously perform traffic object detection, lane line detection, and drivable area segmentation. First, a multi-stage gated aggregation network (MogaNet) is employed during the feature extraction process to enhance contextual interaction by improving diversity in the channel dimension, thereby compensating for the limitations of feed-forward neural networks in contextual understanding. Second, to further improve the model’s accuracy in detecting objects of various scales, a restructured weighted bidirectional feature pyramid network (BiFPN) is introduced, optimizing cross-level information fusion and enabling the model to handle object detection at different scales more accurately. Finally, the model is equipped with one detection head and two segmentation heads to achieve efficient multi-task environmental perception, ensuring the simultaneous execution of multiple tasks. The experimental results on the BDD100K dataset demonstrate that the model achieves a mean average precision (mAP50) of 81.4% in object detection, an Intersection over Union (IoU) of 28.9% in lane detection, and a mean Intersection over Union (mIoU) of 92.6% in drivable area segmentation. The tests conducted in real-world scenarios show that the model performs effectively, significantly enhancing environmental perception in autonomous driving and laying a solid foundation for safer and more reliable autonomous driving systems.
Funder
15th Graduate Education Innovative Fund of Wuhan Institute of Technology, China
Reference31 articles.
1. Faster R-CNN: Towards real-time object detection with region proposal networks;Ren;IEEE Trans. Pattern Anal. Mach. Intell.,2016 2. Chen, Q., Wang, Y., Yang, T., Zhang, X., Cheng, J., and Sun, J. (2021, January 20–25). You only look one-level feature. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition 2021, Nashville, TN, USA. 3. Zhang, S., Wang, X., Wang, J., Pang, J., Lyu, C., Zhang, W., Luo, P., and Chen, K. (2023, January 17–24). Dense distinct query for end-to-end object detection. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition 2023, Vancouver, BC, Canada. 4. Ultra fast deep lane detection with hybrid anchor driven ordinal classification;Qin;IEEE Trans. Pattern Anal. Mach. Intell.,2022 5. Wang, J., Ma, Y., Huang, S., Hui, T., Wang, F., Qian, C., and Zhang, T. (2022, January 18–24). A keypoint-based global association network for lane detection. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition 2022, New Orleans, LA, USA.
|
|