Vision-Based Situational Graphs Exploiting Fiducial Markers for the Integration of Semantic Entities
Author:
Tourani Ali12ORCID, Bavle Hriday1ORCID, Avşar Deniz Işınsu23, Sanchez-Lopez Jose Luis1ORCID, Munoz-Salinas Rafael4ORCID, Voos Holger12ORCID
Affiliation:
1. Interdisciplinary Centre for Security, Reliability, and Trust (SnT), University of Luxembourg, L-1855 Luxembourg, Luxembourg 2. Institute for Advanced Studies, University of Luxembourg, L-4365 Esch-sur-Alzette, Luxembourg 3. Department of Physics & Materials Science, University of Luxembourg, L-1511 Luxembourg, Luxembourg 4. Department of Computer Science and Numerical Analysis, Rabanales Campus, University of Córdoba, 14071 Córdoba, Spain
Abstract
Situational Graphs (S-Graphs) merge geometric models of the environment generated by Simultaneous Localization and Mapping (SLAM) approaches with 3D scene graphs into a multi-layered jointly optimizable factor graph. As an advantage, S-Graphs not only offer a more comprehensive robotic situational awareness by combining geometric maps with diverse, hierarchically organized semantic entities and their topological relationships within one graph, but they also lead to improved performance of localization and mapping on the SLAM level by exploiting semantic information. In this paper, we introduce a vision-based version of S-Graphs where a conventional Visual SLAM (VSLAM) system is used for low-level feature tracking and mapping. In addition, the framework exploits the potential of fiducial markers (both visible and our recently introduced transparent or fully invisible markers) to encode comprehensive information about environments and the objects within them. The markers aid in identifying and mapping structural-level semantic entities, including walls and doors in the environment, with reliable poses in the global reference, subsequently establishing meaningful associations with higher-level entities, including corridors and rooms. However, in addition to including semantic entities, the semantic and geometric constraints imposed by the fiducial markers are also utilized to improve the reconstructed map’s quality and reduce localization errors. Experimental results on a real-world dataset collected using legged robots show that our framework excels in crafting a richer, multi-layered hierarchical map and enhances robot pose accuracy at the same time.
Funder
Luxembourg National Research Fund University of Luxembourg
Reference34 articles.
1. Macario Barros, A., Michel, M., Moline, Y., Corre, G., and Carrel, F. (2022). A comprehensive survey of visual slam algorithms. Robotics, 11. 2. Rosinol, A., Abate, M., Chang, Y., and Carlone, L. (2020). Kimera: An Open-Source Library for Real-Time Metric-Semantic Localization and Mapping. arXiv. 3. Armeni, I., He, Z.Y., Gwak, J., Zamir, A.R., Fischer, M., Malik, J., and Savarese, S. (November, January 27). 3D Scene Graph: A Structure for Unified Semantics, 3D Space, and Camera. Proceedings of the IEEE International Conference on Computer Vision, Seoul, Republic of Korea. 4. Rosinol, A., Gupta, A., Abate, M., Shi, J., and Carlone, L. (2020). 3D Dynamic Scene Graphs: Actionable Spatial Perception with Places, Objects, and Humans. arXiv. 5. Hughes, N., Chang, Y., and Carlone, L. (2022). Hydra: A Real-time Spatial Perception System for 3D Scene Graph Construction and Optimization. arXiv.
Cited by
1 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
|
|