Abstract
AbstractFully convolutional neural network has shown advantages in the salient object detection by using the RGB or RGB-D images. However, there is an object-part dilemma since most fully convolutional neural network inevitably leads to an incomplete segmentation of the salient object. Although the capsule network is capable of recognizing a complete object, it is highly computational demand and time consuming. In this paper, we propose a novel convolutional capsule network based on feature extraction and integration for dealing with the object-part relationship, with less computation demand. First and foremost, RGB features are extracted and integrated by using the VGG backbone and feature extraction module. Then, these features, integrating with depth images by using feature depth module, are upsampled progressively to produce a feature map. In the next step, the feature map is fed into the feature-integrated convolutional capsule network to explore the object-part relationship. The proposed capsule network extracts object-part information by using convolutional capsules with locally-connected routing and predicts the final salient map based on the deconvolutional capsules. Experimental results on four RGB-D benchmark datasets show that our proposed method outperforms 23 state-of-the-art algorithms.
Funder
National Natural Science Foundation of China
Publisher
Springer Science and Business Media LLC
Reference51 articles.
1. Fan, D. P., Zhai Y, Ali, B. et al. BBS-Net: RGB-D Salient Object Detection with a Bifurcated Backbone Strategy Network. https://arxiv.org/abs/02713 (2007).
2. Fu, K. et al. Siamese network for RGB-D salient object detection and beyond. IEEE Trans. Pattern Anal. Mach. Intell. 44(9), 5541–5559 (2021).
3. Pang, Y. W., Zhang, L. H., Zhao, X. Q. et al. Hierarchical dynamic filtering network for RGB-D salient object detection. In European Conference on Computer Vision 235–252 (2020).
4. Zhang, J., Fan, D. P., Dai Y. C. et al. UC-Net: Uncertainty inspired RGB-D saliency detection via conditional variational autoencoders. In IEEE Conference on Computer Vision and Pattern Recognition (2020).
5. Chen, H. & Li, Y. Progressively complementarity-aware fusion network for RGB-D salient object detection. In IEEE Conf. Comput. Vis. Pattern Recog. 3051–3060 (2018).
Cited by
2 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献