Probabilistic Modeling of Multicamera Interference for Time-of-Flight Sensors
Author:
Rodriguez Bryan1ORCID, Zhang Xinxiang1ORCID, Rajan Dinesh1
Affiliation:
1. Department of Electrical and Computer Engineering, Lyle School of Engineering, Southern Methodist University, Dallas, TX 75205, USA
Abstract
The behavior of multicamera interference in 3D images (e.g., depth maps), which is based on infrared (IR) light, is not well understood. In 3D images, when multicamera interference is present, there is an increase in the amount of zero-value pixels, resulting in a loss of depth information. In this work, we demonstrate a framework for synthetically generating direct and indirect multicamera interference using a combination of a probabilistic model and ray tracing. Our mathematical model predicts the locations and probabilities of zero-value pixels in depth maps that contain multicamera interference. Our model accurately predicts where depth information may be lost in a depth map when multicamera interference is present. We compare the proposed synthetic 3D interference images with controlled 3D interference images captured in our laboratory. The proposed framework achieves an average root mean square error (RMSE) of 0.0625, an average peak signal-to-noise ratio (PSNR) of 24.1277 dB, and an average structural similarity index measure (SSIM) of 0.9007 for predicting direct multicamera interference, and an average RMSE of 0.0312, an average PSNR of 26.2280 dB, and an average SSIM of 0.9064 for predicting indirect multicamera interference. The proposed framework can be used to develop and test interference mitigation techniques that will be crucial for the successful proliferation of these devices.
Subject
Electrical and Electronic Engineering,Biochemistry,Instrumentation,Atomic and Molecular Physics, and Optics,Analytical Chemistry
Reference45 articles.
1. SAFER vehicle inspection: A multimodal robotic sensing platform;Page;Unmanned Ground Vehicle Technology VI, Proceedings of the Defense and Security, Orlando, FL, USA, 12–16 April 2004,2004 2. Chen, C., Yang, B., Song, S., Tian, M., Li, J., Dai, W., and Fang, L. (2018). Calibrate Multiple Consumer RGB-D Cameras for Low-Cost and Efficient 3D Indoor Mapping. Remote Sens., 10. 3. Rodriguez, B., Zhang, X., and Rajan, D. (2021, January 25–27). Synthetically Generating Motion Blur in a Depth Map from Time-of-Flight Sensors. Proceedings of the 2021 17th International Conference on Machine Vision and Applications (MVA), Aichi, Japan. 4. Rodriguez, B., Zhang, X., and Rajan, D. (2022). Probabilistic Modeling of Motion Blur for Time-of-Flight Sensors. Sensors, 22. 5. Paredes, J.A., Álvarez, F.J., Aguilera, T., and Villadangos, J.M. (2018). 3D indoor positioning of UAVs with spread spectrum ultrasound and time-of-flight cameras. Sensors, 18.
Cited by
1 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
|
|