DiffuPrompter: Pixel-Level Automatic Annotation for High-Resolution Remote Sensing Images with Foundation Models
-
Published:2024-06-02
Issue:11
Volume:16
Page:2004
-
ISSN:2072-4292
-
Container-title:Remote Sensing
-
language:en
-
Short-container-title:Remote Sensing
Author:
Li Huadong12ORCID, Wei Ying1, Peng Han2ORCID, Zhang Wei2
Affiliation:
1. College of Information Science and Engineering, Northeastern University, Shenyang 110819, China 2. Peng Cheng Laboratory, Shenzhen 518055, China
Abstract
Instance segmentation is pivotal in remote sensing image (RSI) analysis, aiding in many downstream tasks. However, annotating images with pixel-wise annotations is time-consuming and laborious. Despite some progress in automatic annotation, the performance of existing methods still needs improvement due to the high precision requirements for pixel-level annotation and the complexity of RSIs. With the support of large-scale data, some foundational models have made significant progress in semantic understanding and generalization capabilities. In this paper, we delve deep into the potential of the foundational models in automatic annotation and propose a training-free automatic annotation method called DiffuPrompter, achieving pixel-level automatic annotation of RSIs. Extensive experimental results indicate that the proposed method can provide reliable pseudo-labels, significantly reducing the annotation costs of the segmentation task. Additionally, the cross-domain validation experiments confirm the powerful effectiveness of large-scale pseudo-data in improving model generalization performance.
Funder
National Nature Science Foundation of China Open Project Program Foundation of the Key Laboratory of Opto-Electronics Information Processing, Chinese Academy of Sciences
Reference45 articles.
1. Chen, K., Zou, Z., and Shi, Z. (2021). Building extraction from remote sensing images with sparse token transformers. Remote Sens., 13. 2. A survey and analysis on automatic image annotation;Cheng;Pattern Recognit.,2018 3. Wu, T., Huang, J., Gao, G., Wei, X., Wei, X., Luo, X., and Liu, C.H. (2021, January 20–25). Embedded discriminative attention mechanism for weakly supervised semantic segmentation. Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA. 4. Xu, L., Ouyang, W., Bennamoun, M., Boussaid, F., Sohel, F., and Xu, D. (2021, January 10–17). Leveraging auxiliary tasks with affinity learning for weakly supervised semantic segmentation. Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada. 5. Ru, L., Zhan, Y., Yu, B., and Du, B. (2022, January 18–24). Learning affinity from attention: End-to-end weakly-supervised semantic segmentation with transformers. Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA.
|
|