Anthropogenic Object Localization: Evaluation of Broad-Area High-Resolution Imagery Scans Using Deep Learning in Overhead Imagery
Author:
Hurt J. Alex1, Popescu Ilinca2, Davis Curt H.13, Scott Grant J.13
Affiliation:
1. Center for Geospatial Intelligence, University of Missouri, Columbia, MO 65211, USA 2. Department of Geography, Stanford University, Stanford, CA 94305, USA 3. Department of Electrical Engineering and Computer Science, University of Missouri, Columbia, MO 65211, USA
Abstract
Too often, the testing and evaluation of object detection, as well as the classification techniques for high-resolution remote sensing imagery, are confined to clean, discretely partitioned datasets, i.e., the closed-world model. In recent years, the performance on a number of benchmark datasets has exceeded 99% when evaluated using cross-validation techniques. However, real-world remote sensing data are truly big data, which often exceed billions of pixels. Therefore, one of the greatest challenges regarding the evaluation of machine learning models taken out of the clean laboratory setting and into the real world is the difficulty of measuring performance. It is necessary to evaluate these models on a grander scale, namely, tens of thousands of square kilometers, where it is intractable to the ground truth and the ever-changing anthropogenic surface of Earth. The ultimate goal of computer vision model development for automated analysis and broad area search and discovery is to augment and assist humans, specifically human–machine teaming for real-world tasks. In this research, various models have been trained using object classes from benchmark datasets such as UC Merced, PatternNet, RESISC-45, and MDSv2. We detail techniques to scan broad swaths of the Earth with deep convolutional neural networks. We present algorithms for localizing object detection results, as well as a methodology for the evaluation of the results of broad-area scans. Our research explores the challenges of transitioning these models out of the training–validation laboratory setting and into the real-world application domain. We show a scalable approach to leverage state-of-the-art deep convolutional neural networks for the search, detection, and annotation of objects within large swaths of imagery, with the ultimate goal of providing a methodology for evaluating object detection machine learning models in real-world scenarios.
Subject
Electrical and Electronic Engineering,Biochemistry,Instrumentation,Atomic and Molecular Physics, and Optics,Analytical Chemistry
Reference37 articles.
1. Yang, Y., and Newsam, S. (2010, January 2–5). Bag-of-Visual-Words and Spatial Extensions for Land-Use Classification. Proceedings of the ACM SIGSPATIAL International Conference on Advances in Geographic Information Systems (ACM GIS), San Jose, CA, USA. 2. Satellite image classification via two-layer sparse coding with biased image representation;Dai;IEEE Geosci. Remote Sens. Lett.,2011 3. Zhou, W., Newsam, S., Li, C., and Shao, Z. (2017). PatternNet: A Benchmark Dataset for Performance Evaluation of Remote Sensing Image Retrieval. arXiv. 4. Li, H., Tao, C., Wu, Z., Chen, J., Gong, J., and Deng, M. (2017). RSI-CB: A Large Scale Remote Sensing Image Classification Benchmark via Crowdsource Data. arXiv. 5. Xia, G., Hu, J., Hu, F., Shi, B., Bai, X., Zhong, Y., and Zhang, L. (2016). AID: A Benchmark Dataset for Performance Evaluation of Aerial Scene Classification. arXiv.
|
|