Cross-pollination of knowledge for object detection in domain adaptation for industrial automation
-
Published:2024-09-09
Issue:
Volume:
Page:
-
ISSN:2366-5971
-
Container-title:International Journal of Intelligent Robotics and Applications
-
language:en
-
Short-container-title:Int J Intell Robot Appl
Author:
Rehman Anwar Ur,Gallo Ignazio
Abstract
AbstractArtificial Intelligence is revolutionizing industries by enhancing efficiency through real-time Object Detection (OD) applications. Utilizing advanced computer vision techniques, OD systems automate processes, analyze complex visual data, and facilitate data-driven decisions, thus increasing productivity. Domain Adaptation for OD has recently gained prominence for its ability to recognize target objects without annotations. Innovative approaches that merge traditional cross-disciplinary domain modeling with cutting-edge deep learning have become essential in addressing complex AI challenges in real-time scenarios. Unlike traditional methods, this study proposes a novel, effective Cross-Pollination of Knowledge (CPK) strategy for domain adaptation inspired by botanical processes. The CPK approach involves merging target samples with source samples at the input stage. By incorporating a random and unique selection of a few target samples, the merging process enhances object detection results efficiently in domain adaptation, supporting detectors in aligning and generalizing features with the source domain. Additionally, this work presents the new Planeat digit recognition dataset, which includes 231 images. To ensure robust comparison, we employ a self-supervised Domain Adaptation (UDA) method that simultaneously trains target and source domains using unsupervised techniques. UDA method leverages target data to identify high-confidence regions, which are then cropped and augmented, adapting UDA for effective OD. The proposed CPK approach significantly outperforms existing UDA techniques, improving mean Average Precision (mAP) by 10.9% through rigorous testing on five diverse datasets across different conditions- cross-weather, cross-camera, and synthetic-to-real. Our code is publicly available https://github.com/anwaar0/CPK-Object-Detection
Funder
Università degli Studi dell'Insubria
Publisher
Springer Science and Business Media LLC
Reference68 articles.
1. Alvarez, J.M., Gevers, T., LeCun, Y., Lopez, A.M.: Road scene segmentation from a single image. In: Computer Vision–ECCV 2012: 12th European Conference on Computer Vision, Florence, Italy, October 7-13, 2012, Proceedings, Part VII 12, pp. 376–389 (2012). Springer 2. Bataduwaarachchi, S.D., Sattarzadeh, A.R., Stewart, M., Ashcroft, B., Morrison, A., North, S., et al.: Towards autonomous cross-pollination: Portable multi-classification system for in situ growth monitoring of tomato flowers. Smart Agr. Technol. 4, 100205 (2023) 3. Ben-David, S., Blitzer, J., Crammer, K., Kulesza, A., Pereira, F., Vaughan, J.W.: A theory of learning from different domains. Mach. Learn. 79, 151–175 (2010) 4. Bochkovskiy, A., Wang, C.-Y., Liao, H.-Y.M.: Yolov4: Optimal speed and accuracy of object detection. arXiv preprint arXiv:2004.10934 (2020) 5. Cordts, M., Omran, M., Ramos, S., Rehfeld, T., Enzweiler, M., Benenson, R., Franke, U., Roth, S., Schiele, B.: The cityscapes dataset for semantic urban scene understanding. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3213–3223 (2016)
|
|