A Cross-Level Iterative Subtraction Network for Camouflaged Object Detection
-
Published:2024-09-09
Issue:17
Volume:14
Page:8063
-
ISSN:2076-3417
-
Container-title:Applied Sciences
-
language:en
-
Short-container-title:Applied Sciences
Author:
Hu Tongtong1, Zhang Chao2, Lyu Xin13ORCID, Sun Xiaowen4, Chen Shangjing1ORCID, Zeng Tao1, Chen Jiale1
Affiliation:
1. College of Computer Science and Software Engineering, Hohai University, Nanjing 211100, China 2. Information Center, Ministry of Water Resources, Beijing 100053, China 3. Key Laboratory of Water Big Data Technology of Ministry of Water Resources, Hohai University, Nanjing 211100, China 4. Water Resources Service Center of Jiangsu Province, Nanjing 210029, China
Abstract
Camouflaged object detection (COD) is a challenging task, aimed at segmenting objects that are similar in color and texture to their background. Sufficient multi-scale feature fusion is crucial for accurately segmenting object regions. However, most methods usually focus on information compensation, overlooking the difference between features, which is important for distinguishing the object from the background. To this end, we propose the cross-level iterative subtraction network (CISNet), which integrates information from cross-layer features and enhances details through iteration mechanisms. CISNet involves a cross-level iterative structure (CIS) for feature complementarity, where texture information is used to enrich high-level features and semantic information is used to enhance low-level features. In particular, we present a multi-scale strip convolution subtraction (MSCSub) module within CIS to extract difference information between cross-level features and fuse multi-scale features, which improves the feature representation and guides accurate segmentation. Furthermore, an enhanced guided attention (EGA) module is presented to refine features by deeply mining local context information and capturing a broader range of relationships between different feature maps in a top-down manner. Extensive experiments conducted on four benchmark datasets demonstrate that our model outperforms the state-of-the-art COD models in all evaluation metrics.
Funder
National Key Research and Development Program of China Excellent Post-doctoral Program of Jiangsu Province Fundamental Research Funds for the Central Universities
Reference78 articles.
1. Dong, B., Wang, W., Fan, D.P., Li, J., Fu, H., and Shao, L. (2021). Polyp-pvt: Polyp segmentation with pyramid vision transformers. arXiv. 2. Inf-net: Automatic covid-19 lung infection segmentation from ct images;Fan;IEEE Trans. Med. Imaging,2020 3. Fan, D.P., Ji, G.P., Zhou, T., Chen, G., Fu, H., Shen, J., and Shao, L. (2020, January 4–8). Pranet: Parallel reverse attention network for polyp segmentation. Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Lima, Peru. 4. Medical Image Segmentation Based on Object Detection;Deng;J. Univ. Electron. Sci. Technol. China,2023 5. Xie, E., Wang, W., Wang, W., Ding, M., Shen, C., and Luo, P. (2020, January 23–28). Segmenting transparent objects in the wild. Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK. Proceedings, Part XIII 16.
|
|