Learning Domain-Adaptive Landmark Detection-Based Self-Supervised Video Synchronization for Remote Sensing Panorama
-
Published:2023-02-09
Issue:4
Volume:15
Page:953
-
ISSN:2072-4292
-
Container-title:Remote Sensing
-
language:en
-
Short-container-title:Remote Sensing
Author:
Mei Ling1ORCID, He Yizhuo2ORCID, Fishani Farnoosh Javadi3, Yu Yaowen4ORCID, Zhang Lijun4, Rhodin Helge3
Affiliation:
1. School of Information Science and Engineering, Wuhan University of Science and Technology, Wuhan 430081, China 2. School of Computer Science, Carnegie Mellon University (CMU), Pittsburgh, PA 15213, USA 3. Department of Computer Science, University of British Columbia (UBC), Vancouver, BC V6T 1Z4, Canada 4. School of Artificial Intelligence and Automation, Huazhong University of Science and Technology, Wuhan 430074, China
Abstract
The synchronization of videos is an essential pre-processing step for multi-view reconstruction such as the image mosaic by UAV remote sensing; it is often solved with hardware solutions in motion capture studios. However, traditional synchronization setups rely on manual interventions or software solutions and only fit for a particular domain of motions. In this paper, we propose a self-supervised video synchronization algorithm that attains high accuracy in diverse scenarios without cumbersome manual intervention. At the core is a motion-based video synchronization algorithm that infers temporal offsets from the trajectories of moving objects in the videos. It is complemented by a self-supervised scene decomposition algorithm that detects common parts and their motion tracks in two or more videos, without requiring any manual positional supervision. We evaluate our approach on three different datasets, including the motion of humans, animals, and simulated objects, and use it to build the view panorama of the remote sensing field. All experiments demonstrate that the proposed location-based synchronization is more effective compared to the state-of-the-art methods, and our self-supervised inference approaches the accuracy of supervised solutions, while being much easier to adapt to a new target domain.
Funder
Fundamental Research Funds for the Central Universities, HUST International Program Award for Young Talent Scientific Research People of Guangdong Province
Subject
General Earth and Planetary Sciences
Reference72 articles.
1. Aires, A.S., Marques Junior, A., Zanotta, D.C., Spigolon, A.L.D., Veronez, M.R., and Gonzaga, L. (2022). Digital Outcrop Model Generation from Hybrid UAV and Panoramic Imaging Systems. Remote Sens., 14. 2. Zhang, Y., Mei, X., Ma, Y., Jiang, X., Peng, Z., and Huang, J. (2022). Hyperspectral Panoramic Image Stitching Using Robust Matching and Adaptive Bundle Adjustment. Remote Sens., 14. 3. Han, P., Ma, C., Chen, J., Chen, L., Bu, S., Xu, S., Zhao, Y., Zhang, C., and Hagino, T. (2022). Fast Tree Detection and Counting on UAVs for Sequential Aerial Images with Generating Orthophoto Mosaicing. Remote Sens., 14. 4. Hwang, Y.S., Schlüter, S., Park, S.I., and Um, J.S. (2021). Comparative evaluation of mapping accuracy between UAV video versus photo mosaic for the scattered urban photovoltaic panel. Remote Sens., 13. 5. Wandt, B., Little, J.J., and Rhodin, H. (2022, January 18–24). ElePose: Unsupervised 3D Human Pose Estimation by Predicting Camera Elevation and Learning Normalizing Flows on 2D Poses. Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA.
|
|