MineSDS: A Unified Framework for Small Object Detection and Drivable Area Segmentation for Open-Pit Mining Scenario
Author:
Liu Yong1, Li Cheng1, Huang Jiade1, Gao Ming2
Affiliation:
1. Zhuzhou CRRC Times Electric Co., Ltd., Zhuzhou 412001, China 2. State Key Laboratory of Advanced Design and Manufacturing for Vehicle Body, College of Mechanical and Vehicle Engineering, Hunan University, Changsha 410082, China
Abstract
To tackle the challenges posed by dense small objects and fuzzy boundaries on unstructured roads in the mining scenario, we proposed an end-to-end small object detection and drivable area segmentation framework for open-pit mining. We employed a convolutional network backbone as a feature extractor for both two tasks, as multi-task learning yielded promising results in autonomous driving perception. To address small object detection, we introduced a lightweight attention module that allowed our network to focus more on the spatial and channel dimensions of small objects without impeding inference time. We also used a convolutional block attention module in the drivable area segmentation subnetwork, which assigned more weight to road boundaries to improve feature mapping capabilities. Furthermore, to improve our network perception accuracy of both tasks, we used weighted summation when designing the loss function. We validated the effectiveness of our approach by testing it on pre-collected mining data which were called Minescape. Our detection results on the Minescape dataset showed 87.8% mAP index, which was 9.3% higher than state-of-the-art algorithms. Our segmentation results surpassed the comparison algorithm by 1 percent in MIoU index. Our experimental results demonstrated that our approach achieves competitive performance.
Subject
Electrical and Electronic Engineering,Biochemistry,Instrumentation,Atomic and Molecular Physics, and Optics,Analytical Chemistry
Reference38 articles.
1. Balasubramaniam, A., and Pasricha, S. (2023). Object Detection in Autonomous Vehicles: Status and Open Challenges. arXiv. 2. Li, Y., Li, Z., Teng, S., Zhang, Y., Zhou, Y., Zhu, Y., Cao, D., Tian, B., Ai, Y., and Xuanyuan, Z. (2022, January 19–20). AutoMine: An Unmanned Mine Dataset. Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA. 3. Wang, C.-Y., Bochkovskiy, A., and Liao, H.-Y.M. (2023). YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. arXiv. 4. Ferrari, V., Hebert, M., Sminchisescu, C., and Weiss, Y. (2018, January 8–14). CornerNet: Detecting Objects as Paired Keypoints. Proceedings of the Computer Vision—ECCV, Munich, Germany. 5. Fleet, D., Pajdla, T., Schiele, B., and Tuytelaars, T. (2014, January 6–12). Microsoft COCO: Common Objects in Context. Proceedings of the Computer Vision—ECCV 2014, Zurich, Switzerland.
Cited by
1 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
|
|