Author:
Doğan Fethiye Irmak,Melsión Gaspar I.,Leite Iolanda
Abstract
For effective human-robot collaboration, it is crucial for robots to understand requests from users perceiving the three-dimensional space and ask reasonable follow-up questions when there are ambiguities. While comprehending the users’ object descriptions in the requests, existing studies have focused on this challenge for limited object categories that can be detected or localized with existing object detection and localization modules. Further, they have mostly focused on comprehending the object descriptions using flat RGB images without considering the depth dimension. On the other hand, in the wild, it is impossible to limit the object categories that can be encountered during the interaction, and 3-dimensional space perception that includes depth information is fundamental in successful task completion. To understand described objects and resolve ambiguities in the wild, for the first time, we suggest a method leveraging explainability. Our method focuses on the active areas of an RGB scene to find the described objects without putting the previous constraints on object categories and natural language instructions. We further improve our method to identify the described objects considering depth dimension. We evaluate our method in varied real-world images and observe that the regions suggested by our method can help resolve ambiguities. When we compare our method with a state-of-the-art baseline, we show that our method performs better in scenes with ambiguous objects which cannot be recognized by existing object detectors. We also show that using depth features significantly improves performance in scenes where depth data is critical to disambiguate the objects and across our evaluation dataset that contains objects that can be specified with and without the depth dimension.
Funder
Vetenskapsrådet
Stiftelsen för Strategisk Forskning
NordForsk
Kungliga Tekniska Högskolan
Knut och Alice Wallenbergs Stiftelse
Subject
Artificial Intelligence,Computer Science Applications
Reference67 articles.
1. Trends and trajectories for explainable, accountable and intelligible systems: An HCI research agenda;Abdul,2018
2. Referit3d: Neural listeners for fine-grained 3d object identification in real-world scenes;Achlioptas,2020
3. Peeking inside the black-box: A survey on explainable artificial intelligence (XAI);Adadi;IEEE Access,2018
4. Interactive natural language technology for explainable artificial intelligence;Alonso,2021
5. Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI;Barredo Arrieta;Inf. Fusion,2020
Cited by
4 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
1. Deep learning framework with Local Sparse Transformer for construction worker detection in 3D with LiDAR;Computer-Aided Civil and Infrastructure Engineering;2024-05-26
2. Human-in-the-loop error detection in an object organization task with a social robot;Frontiers in Robotics and AI;2024-04-16
3. Explainability for Human-Robot Collaboration;Companion of the 2024 ACM/IEEE International Conference on Human-Robot Interaction;2024-03-11
4. Semantic Scene Understanding for Human-Robot Interaction;Companion of the 2023 ACM/IEEE International Conference on Human-Robot Interaction;2023-03-13