Reinforcement Learning and Genetic Algorithm-Based Network Module for Camera-LiDAR Detection
-
Published:2024-06-22
Issue:13
Volume:16
Page:2287
-
ISSN:2072-4292
-
Container-title:Remote Sensing
-
language:en
-
Short-container-title:Remote Sensing
Author:
Kim Taek-Lim1ORCID, Park Tae-Hyoung2
Affiliation:
1. Department of Control and Robot Engineering, Chungbuk National University, Cheongju 28644, Republic of Korea 2. Department of Intelligent Systems & Robotics, Chungbuk National University, Cheongju 28644, Republic of Korea
Abstract
Cameras and LiDAR sensors have been used in sensor fusion for robust object detection in autonomous driving. Object detection networks for autonomous driving are often trained again by adding or changing datasets aimed at robust performance. Repeat training is necessary to develop an efficient network module. Existing efficient network module development changes to hand design and requires much module design experience. For this, a neural architecture search was designed, but it takes much time and requires optimizing the design process. To solve this problem, we propose a two-stage optimization method for the offspring generation process in a neural architecture search based on reinforcement learning. In addition, we propose utilizing two split datasets to solve the fast convergence problem as the objective function of the genetic algorithm: source data (daytime, sunny) and target data (day/night, adversary weather). The proposed method is an efficient module generation method requiring less time than the NSGA-NET. We confirmed the performance improvement and the convergence speed reduction using the Dense dataset. Through experiments, it was proven that the proposed method generated an efficient module.
Funder
Institute of Information & Communications Technology Planning & Evaluation
Reference62 articles.
1. He, K., Zhang, X., Ren, S., and Sun, J. (26–1, January 26). Deep residual learning for image recognition. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA. 2. Liu, W., Anguelov, D., Erhan, D., Szegedy, C., Reed, S., Fu, C.Y., and Berg, A.C. (2016, January 11–14). Ssd: Single shot multibox detector. Proceedings of the Computer Vision–ECCV (ECCV), Amsterdam, The Netherlands. 3. Redmon, J., and Farhadi, A. (2018). Yolov3: An incremental improvement. arXiv. 4. Bijelic, M., Gruber, T., Mannan, F., Kraus, F., Ritter, W., Dietmayer, K., and Heide, F. (2020, January 16–18). Seeing through fog without seeing fog: Deep multimodal sensor fusion in unseen adverse weather. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA. 5. Multi-modal multi-channel target speech separation;Gu;IEEE J. Sel. Top. Signal Process.,2020
|
|