Affiliation:
1. Key Laboratory of Technology in Geo-Spatial Information Processing and Application System, Aerospace Information Research Institute, Chinese Academy of Sciences, Beijing 100190, China
2. Aerospace Information Research Institute, Chinese Academy of Sciences, Beijing 100094, China
3. School of Electronic, Electrical and Communication Engineering, University of Chinese Academy of Sciences, Beijing 101408, China
Abstract
Instance segmentation of remote sensing images can not only provide object-level positioning information but also provide pixel-level positioning information. This pixel-level information annotation has a wide range of uses in the field of remote sensing, and it is of great value for environmental detection and resource management. Because optical images generally have complex terrain environments and changeable object shapes, SAR images are affected by complex scattering phenomena, and the mask quality obtained by the traditional instance segmentation method used in remote sensing images is not high. Therefore, it is a challenging task to improve the mask quality of instance segmentation in remote sensing images. Since the traditional two-stage instance segmentation method consists of backbone, neck, bbox head, and mask head, the final mask quality depends on the product of all front-end work quality. Therefore, we consider the difficulty of optical and SAR images to bring instance segmentation to the targeted improvement of the neck, bbox head, and mask head, and we propose the Context-Aggregated and SAM-Guided Network (CSNet). In this network, the plain feature fusion pyramid network (PFFPN) can generate a pyramid for the plain feature and provide a feature map of the appropriate instance scale for detection and segmentation. The network also includes a context aggregation bbox head (CABH), which uses the context information and instance information around the instance to solve the problem of missed detection and false detection in detection. The network also has a SAM-Guided mask head (SGMH), which learns by using SAM as a teacher, and uses the knowledge learned to improve the edge of the mask. Experimental results show that CSNet significantly improves the quality of masks generated under optical and SAR images, and CSNet achieves 5.1% and 3.2% AP increments compared with other SOTA models.
Reference71 articles.
1. Gao, J., Zhang, B., Wu, Y., and Guo, C. (2022, January 25). Building Extraction from High Resolution Remote Sensing Images Based on Improved Mask R-CNN. Proceedings of the 2022 4th International Conference on Robotics and Computer Vision (ICRCV), Wuhan, China.
2. DF-Mask R-CNN: Direction Field-Based Optimized Instance Segmentation Network for Building Instance Extraction;Zhang;IEEE Geosci. Remote Sens. Lett.,2023
3. Multiscale U-Shaped CNN Building Instance Extraction Framework with Edge Constraint for High-Spatial-Resolution Remote Sensing Imagery;Liu;IEEE Trans. Geosci. Remote Sens.,2021
4. Tian, T., Gao, P., Pan, Z., Li, H., and Wang, L. (2020, January 26). Inshore Ship Detection Based on Multi-Information Fusion Network and Instance Segmentation. Proceedings of the IGARSS 2020—2020 IEEE International Geoscience and Remote Sensing Symposium, Waikoloa, HI, USA.
5. LFG-Net: Low-Level Feature Guided Network for Precise Ship Instance Segmentation in SAR Images;Wei;IEEE Trans. Geosci. Remote Sens.,2022