Affiliation:
1. Changchun Institute of Optics, Fine Mechanics and Physics, Chinese Academy of Sciences, Changchun 130033, China
2. University of Chinese Academy of Sciences, Beijing 100049, China
Abstract
The complementary characteristics of SAR and optical images are beneficial in improving the accuracy of land cover classification. Deep learning-based models have achieved some notable results. However, how to effectively extract and fuse the unique features of multi-modal images for pixel-level classification remains challenging. In this article, a two-branch supervised semantic segmentation framework without any pretrained backbone is proposed. Specifically, a novel symmetric attention module is designed with improved strip pooling. The multiple long receptive fields can better perceive irregular objects and obtain more anisotropic contextual information. Meanwhile, to solve the semantic absence and inconsistency of different modalities, we construct a multi-scale fusion module, which is composed of atrous spatial pyramid pooling, varisized convolutions and skip connections. A joint loss function is introduced to constrain the backpropagation and reduce the impact of class imbalance. Validation experiments were implemented on the DFC2020 and WHU-OPT-SAR datasets. The proposed model achieved the best quantitative values on the metrics of OA, Kappa and mIoU, and its class accuracy was also excellent. It is worth mentioning that the number of parameters and the computational complexity of the method are relatively low. The adaptability of the model was verified on RGB–thermal segmentation task.
Funder
National Key R&D Program of China
Reference49 articles.
1. More Diverse Means Better: Multimodal Deep Learning Meets Remote-Sensing Imagery Classification;Hong;IEEE Trans. Geosci. Remote Sens.,2021
2. Li, X., Zhang, G., Cui, H., Hou, S., Wang, S., Li, X., Chen, Y., Li, Z., and Zhang, L. (2022). MCANet: A joint semantic segmentation framework of optical and SAR images for land use classification. Int. J. Appl. Earth Obs., 106.
3. Multimodal Bilinear Fusion Network With Second-Order Attention-Based Channel Selection for Land Cover Classification;Li;IEEE J. Sel. Top. Appl. Earth Observ. Remote Sens.,2020
4. Image fusion in the loop of high-level vision tasks: A semantic-aware real-time infrared and visible image fusion network;Tang;Inform. Fusion,2022
5. Collaborative Attention-Based Heterogeneous Gated Fusion Network for Land Cover Classification;Li;IEEE Trans. Geosci. Remote Sens.,2021
Cited by
2 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献