Detection of Novel Objects without Fine-Tuning in Assembly Scenarios by Class-Agnostic Object Detection and Object Re-Identification
-
Published:2024-08-19
Issue:3
Volume:5
Page:373-406
-
ISSN:2673-4052
-
Container-title:Automation
-
language:en
-
Short-container-title:Automation
Author:
Eisenbach Markus1ORCID, Franke Henning1, Franze Erik1, Köhler Mona1ORCID, Aganian Dustin1ORCID, Seichter Daniel1ORCID, Gross Horst-Michael1ORCID
Affiliation:
1. Neuroinformatics and Cognitive Robotics Lab, Ilmenau University of Technology, 98693 Ilmenau, Germany
Abstract
Object detection is a crucial capability of autonomous agents for human–robot collaboration, as it facilitates the identification of the current processing state. In industrial scenarios, it is uncommon to have comprehensive knowledge of all the objects involved in a given task. Furthermore, training during deployment is not a viable option. Consequently, there is a need for a detector that is able to adapt to novel objects during deployment without the necessity of retraining or fine-tuning on novel data. To achieve this, we propose to exploit the ability of discriminative embeddings learned by an object re-identification model to generalize to unknown categories described by a few shots. To do so, we extract object crops with a class-agnostic detector and then compare the object features with the prototypes of the novel objects. Moreover, we demonstrate that the embedding is also effective for predicting regions of interest, which narrows the search space of the class-agnostic detector and, consequently, increases processing speed. The effectiveness of our approach is evaluated in an assembly scenario, wherein the majority of objects belong to categories distinct from those present in the training datasets. Our experiments demonstrate that, in this scenario, our approach outperforms the current best few-shot object-detection approach DE-ViT, which also does not perform fine-tuning on novel data, in terms of both detection capability and inference speed.
Funder
Carl Zeiss Foundation
Reference97 articles.
1. Eisenbach, M., Aganian, D., Köhler, M., Stephan, B., Schroeter, C., and Gross, H.M. (2021, January 23–27). Visual Scene Understanding for Enabling Situation-Aware Cobots. Proceedings of the IEEE International Conference on Automation Science and Engineering (CASE), Lyon, France. 2. Aganian, D., Köhler, M., Baake, S., Eisenbach, M., and Groß, H.M. (2023, January 18–23). How object information improves skeleton-based human action recognition in assembly tasks. Proceedings of the 2023 International Joint Conference on Neural Networks (IJCNN), Gold Coast, Australia. 3. Li, W., Wei, H., Wu, Y., Yang, J., Ruan, Y., Li, Y., and Tang, Y. (2024). TIDE: Test-Time Few-Shot Object Detection. IEEE Trans. Syst. Man Cybern. Syst. 4. Few-shot object detection: A survey;Antonelli;ACM Comput. Surv. (CSUR),2022 5. Köhler, M., Eisenbach, M., and Gross, H.M. (2023). Few-shot object detection: A comprehensive survey. IEEE Trans. Neural Netw. Learn. Syst., 1–21.
|
|